var/home/core/zuul-output/0000755000175000017500000000000015134367774014546 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015134375510015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000240253415134375447020275 0ustar corecore'qikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD >lEڤ펯_ˎ6Ϸ7+%f?長ox[o8W5!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;nt 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'B쿫dIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË?/,z_jx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!frH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{BpY]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&Va%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKd?{WƱPz;| \;_D[T/BI GH8@"t*"9A(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|04SRm+0^PTi-"] O('@BKD6 {NmʐzRj.aQcb^CZ-uvpr CѐٱlGNzIveca=%1Qi F>wTLHUGӃ\sA֎Xpljlv ^tSȻ \cPwίwX"{>9V0ټ_`#U8VdTtD_GU9V ұ{q:ObUi7s )B ۊZlzIA4S#x,T3ѱ ԶJ=rs>Nb: Q6ˌ߉J%.Dl2ȱ%ܱ&6XƟ6qg(USok+Po$lwvmi8W_VT18V =| ub6QWCnY'"*aN08wuSEAVخ m3 o\` sHc# fqT .,ŀU|⦍߶/*~48âF,#[:y_YIpʼn)dk!J'Z5=r&; (y*b*O_ULT.ÔD[%s1,jЅ@k0Ցu֯dtKl$Y5O*GUڇvI`b0ο0~oI`b#FOf_$0!i rS/wvҍ%Eb/Ec|U9F-)L)ŘF`U:VK jeFrԋ7EDYpԽ.D\dNyj荊EEg]bÔF˩ք%EGƶ*NX)Hc(<|q@Oޯr^3>Uf1w;mCja:-1_k٘%VbZ˙#G6 `q+MPU~l!.?I_Pĝ"] rT [eTr؟˰ ]\ h! v˱>5S1px fnk}sRmA>d2UAkؖvlX܇Bz1U_#Xӫ+al H d\k/I,k,ρ|`zR/$@8VU^rcG"E7\qtS:ڝUyy >Vc11*?xYa8U`Jw/AcL~|;yj8TR#s"Q.ϊ/Yrx+u6*27fǪC%+A~*Zآ'ѭnۡ|< a1s\ T5҃FZh?EV"sd!@БU ^p%pO3|B5=2怕nwRqR9~ i±za+HFNi>. EWz:V^&YEs5Ȭ N *7{!fRБBSۘ† Er/IGU}APQT]|XN X]FbKjKdO U6[3TTX)|*H'2U0:VunBl  `5/@ա06VNO8VGON@KgjyK?Wq1egI+ I.*F~L!Gf"LD&U 6tGd#fR*c ^tSLjnKS9 Ȼ \ >lr&}+̼d"I va,Jm_u)d靕َ| Vw85F3Liƙb<;dM-})C?Fw*IJ_3UG'+¨[9| >80\+ xJpΕ`p~mg˗%F Rg(6=/r+%a>w Ohght uЍaRs ^d6GXAf?V_mW puȇ S:tŴvŀU#-*mZ5k5r)_x*8ͼx@(k:_TX%[paRu~}#Ѥr %A%`;MxB[CzR怕#H% }8@*AM.SEhd,rKrʇ)br\+! s1CtӒNc_:F*`Nv;ogQFa2V%ZniE|nZ&-I,t*ώlo Lhnٓ'Xm R ˍ-~ά}hs\5TT%~am.>!LcoJrKmqvez܅E9t6FZXgsreHhlٷ+ [}r:̓?W~e6>0E8`Jq-(ed;W¨:Ä&]䒿e;0:|$Ȃ1L-%;Ƅ{dɱL;V[bp>!n&աIJX1$9;[?- й vRCxKVV+#lj@_RL;IQ8ŢΌXD@Z< (1ZRÜ:OUM/vư{'jYXE4S/8 7: `/ +G\ U>]B2/n2=8) B gJ3bcKo̹ʇ\B~Is 2sO/I!}xV&\b<9$4Nve^آ]$LGF@LjKٕyzH 31Հm-XıUXF|\A-2) ' RG6h?āUŔyj[j_ӂ~ яA弆^bDyzǖQ8`jXbsK?l58,?YP5䜭ve9YFznTEf3Ja\,@2,?WYؾNr<V` =V[oB5!Z\ļǪЎr8@*ucѡv\[|s L-+y{5K@dzp`r"mũɸHNd"yc Pu>x2;W`_VR<aӗ&D<=h-Rר|/r _ǖھcߖ]G@Ն;UQG1 '3Jە Q88ASUȿ!:WѥLf21;d9OU᧯MR3V:<}xXh//T+coY5Ȧ4/m0NE(G2[+G~H'5ipӘ͏O +Px SPp.,?Uv|$kYʈ/K<ݑ.!tB Y"vg|k`?l h3e Mruu=r!-h{D e\y@dY[ vզ(eJ/ie~WՠJ`vU!J)JwdYq%oJ5 ]UJ&\.!$YvV1j^MYTݬ=I.A!R$ S&Be~g Ne ->9J]@<8/-2.o[<%e[:%eGnX!,Ɨ3#4ŸF V^zh9|Y\9Y^Wiؑ7n,^xߕ+1Y*#`Q]\Y"k0 -`p-׋P<Ӥ5i>/q㺽hE^3_wY9_iHJe ˗;\ؾS!VA(la"ﮃwxa`w'Yd2xAǮY~h1~Azx?d4-tF$0KO7u2Opr̓K򗤯iBRNomlCL|ww( <"(-mM׺Z^KM -19w}Ul|6i,gjov_yP4NòiAa+(oD1`#כֿm幞Ɏں;'{innz瘿DH~]y#x/=R /7nudvݝ ohL/?K]_缩Y,YzFf[lTt1)?@۪'48Q@faz]zŴw n),CkYe Ge5Jό}fD<}rm!uӲlm7"^33 [;lOҼ(piz2Qggp":lMU&yȓYg-Ȭ(yQ<4|w?3y)Ça"6 -]q*5"A @l"dLZWq5)xSPrR-& tI>šT (c:tr*dQEzvGwEVWլ}J,|KrWY:H+m5)ulκ0 ^ Y{ʣz%ɟqKK(!DM%xijZgE v(GA/()s–Wr:9j~4 ~xRx&rZYGOgln}ojO+y4ǔ}dq %n?A#z&W"I;S궞y;5w䨍gsxȣe_'M;})h86Q' % alu|I,8~twmAq oc>` {zs-'yݮCBoj$a!Dgqideg OGoi\ zс@LsZ,d)[rN b|.>8ߐ0/8ˤ|6tT# =5\.6x>iچ;sΓ,> PeydBi4o' yI`|!K޼AZ:tFoJ ǧK;j5?*cAoy%.|F$x.yZ$F~)CRPx#$Wa|GlzuAò"|wt*Ûlq$ǣ7mL@dSP Xfg:5l a#r[vdIl>sRdPx,@%d ^4P{(٘^XX!r!Au?sDA_I/]sD%Df ~Eۚ-=վESVfAw]o蜶K2o<n x|eG+aDZϯ$|^!:MBӧ(u稜$ZUƱGvq mui+e8SUh+[{ˈ =h OaH !Ui[jS=$S;M'RT V*LCh}ڜLQQkF>X㨬1xD!QrtSIa=!˸˛ g-W" o( "v0F0JHJ~ˁgaW?m Hk3TVቜxÌH) Yq6Uw,)-*o<<~XF碍Rئp6W{ٵݧR[Q-c ^ (nGdVk64!1Ta 49,(mOs9 xa[izvO3T Alfyp k_ز+%:ASV#$VQ@JB ,}HȈ7IOo^V!Xk@4$Ԟ_[KHJo:qJ [`&)V<wrf50og{;FY˸ր05om5]m䱭%VZY^kF<¤Dr[x`Axe <)γg*)Y",W@A0iwXo}um;ETk`)?"Ep$uOHNyGLVL4^"X:ڇB)!Ug6ΖXozjXXrF x7qDFc[/A4cC75Ԛ:>xS=M.M|h׷zזcp fN ן;5O[uX%\(DNZ7_=$b_.œ(C{;U{5퓤bʣ\rݵf|oVXP1ƈ+ka,j:vWnM]\ЖT8+tז،ٮ4M΢{ fe;әi13ayoeDžDvڛ4{b sVYI|_WQů51$z TcDϔtnLL;6L!-ϒHK3P8dm0kUICt>ڸ~S1?`RyCЁzK3-EC$2kbGO%:eP۲RޡgxC4%}RWp}6{ig zMg՞տξϺa, Rm4 *Ot;yS2W6U2 X9U4[>.$#qUm?X6OvtA{hf9)Kn&B0O1b ,vMD4ەٺu##'0GZyZMhz Dސ'/ILێqP '30t^& 5"8{C=h"D'qIUiRÓ?ڱG:M0 '|9DxaD= 'w65aG Dg6ai[6"\~Thud$@b8 qM2nMv9C9)h7zvzݛ}$agZxƵ_!^&%2ciܤY`Z$'ڒdACJ28* ̴SےH~uCRiwVd(@k&^ E>N"D*̤QDu|t|I?^5{0ڍ{?_< ~#>LtoMIG2X=}J9jys_v^ۣƧA| "ԽzHq @H.a/ NG'/(7`AP|qz 'Egۈ \#(bEn?nE㒏pp?;y}8h P/~f(\GAPwwNIx` |E˙y Og!(}J5$.G @av|>׃ZaA}p 6e:"szbG,  qK]CP~8D!cf*a(1; vEw /+FCF}gw('^[D[w^>"a=&x4" E KloG@8.H; N"ח4WFC 8=Jx4 emBo/mƬVsӇ`~af(ܞXS3#+{aa7gֶY%7dQ'9,grPi?da$8BqM@[rED~OtO;vE(xݾ#gVN#ea.p">^ YV_0P'd <DP"!^?FgsOJ 4 d]# /qFC~>KǡU wM]HPXxhguܾt0bg P[/r&R+(WUkJNJӠ WmC_ V8\r?\;}pu%u#jϐ }qtC;rt5^i&Y}&t5neߠN8jAn!C=bA7_7 c̶Jp4"e!nUY&I*2rph$2u֏p|,`eXZH,5D0Aep fϫւ(/WCTFv4@/5GYI&e1#"*G WTnƂf,Iә1Ԁz¦ <'Rd*9mZ+W*f")uau%Hi5 3gnÑAum@3qHto\}<+[\Z&FбBR|K(lYqLLX)-7е=qquqHSψFܠ4(ԉoN׃z=8a` ${N 6ŤAth_dCa8JhEjFø\hJZhpdmm̾|sN;O$uY6-Q PWRȂTYwK,arH( X`X VWâZOb1pgM.6ިF@~=ld$a,9!зq E~ O&gޝ[%W> PsP=vl-#t &h4X&3urrq;dBeNqB$IiB\5Pxx8We,q(151C,-N{}C:Sks\p2Kf96ቿ2d M*,fIEƠ~%-5XWc9V]ԓ^-jz5 ݩ7\k<\M x~KE9b ܩjuVA­ lP Rr?ۘU>}VeòhaWH_q+ivIڎ&۽Ó<)WX8݃WrA0 vi} / R<7_{G:SA:+k)y-yJUKݘ3ؕW]x<yk>.q\ݵz7b>eQ?|œIǞF^X,Vd%lĦ'` C=F8gǖXAVd:L\jAj^x2 ~PW3[0%X}v?%FU}Uy`KLZklw ||O SآwZ-rͭZc a873f#dY{ &|S*? _5z6a+Њ [Â?8,|3V^wZ'z_X,>< ejR6FŌ{ܷ]9d zH9Vظ.8DuLp& I]iPaKb2B-@ t0V|U{VMvY(m,зC+xТjiCؒA@pMn恻^m,MViks0M7۶7fN8^5N85!K`;MP|(t|7 2kQQ)ö|fh%w|&y53vN:׍|<U4n&#JVbdy^ӐAAr)M|$dz~`'y 6U`TІRNMnWeR|ko e넲-eʞF({$lBou)ۦO#:=Wcmկ;v}L>oݹNZO#z$[joO4BGjoA(_'oA(ߞP4B# [lA=uI:>P[mA=I:?P[lAh= Ih]B\`z4#P󗼸̉ć/BQ8 t}$4/Nx>3AP&.(:FLr-΋gGpO!tbۄ ǝi)պ uSL4ԁR! l3d|HMnyCa: NwO]~(C<k yU)y+|8&ÀM-CpLJ^0([ ~{e/s(ӫE66v`,"CU@ci U?-?O42+A(Y= C* E`WUm5]H5OcS̷1Y??kaV xmhP[I1+dC?WNc*};8Q'*$U<V,NG1tϰn^PB?PiZ]/#4 z$+U_ q\DcS҄b>*[X-<9_h9> RaLd++Z$\uxu|IQ)Cɇ+3E"ެvsX /-YZFS}`T-ۡܛ4 D^CΆ[fu^L?o al 㸙TLs)S&NnKyߣRe$o5btUF4tHhm)\ ,?Jid\vׅ % i\0`;` cc~^P&RIwN-G<|p~Xeg#aj1S Hτ݂^Bzb,*,e V6N]..Lpaͱȶyboj[Y-Y sW}\H9HS\7 Cy=IuB;&,0DW`Aˈ[0' jXX#ZkL[/2$DQwV.%@]H1)Ѱf l(:]iACejew2P 61<A?Ix~Jo ],K@d*TLn>v=+F W{-}iH&ҍfX_lXm [9ܢomU?;MS iUa,#aدPiȎCOej>{x(kDGWao?۰Gc^(q<I$5|GUc(DZlkK?OߓmFLrtfۖ7E/ރY5a1uD>fGQrxkA̠u BjѨ!Q%BDpַYA;eQ^0El2 r$CH e:/":/)$oH )Tz`,*K-]ɪt92P9zX\;,2{ )DW@sVAR¥!fb;AHsqgQ>mBdq4Ч$!{JBk) +4L3z0|n`ڞ []A%{pC{d4:eV!`W(&JĵW;y2xIhlhʜ{\VFse_A/:m֝S8~1isL Ūҕ1#I3"c ש A1_۾֌>r2MN5^4}<(ل%P;$U˽ss2 z)DPGLۺHRqh%DQ\d 9[`mz0UR{TAbͣP ?A̧nnui;(J'䓵9U.s3|>~>}^~=t{+6?=3{vpϽV+nNs$WRx HţyDE5EbMTLIJ4ڱRZƘzoN5FOl$V jܖ MtPh}@ @P&Eɬps1E#Tќ_,Q02tǿ;‰Vkm+Կ _}Cʷn1Ћtwk[jX{~9QWZF O`qGlB74Ȫ\y}t%mCmr ޮ4'd.=ގ{ԑ@XUw_Gr0>׈kV[MLqwR4%(ies7_( }cYt{\bOu nqRRvpW| a0Vk>hnIPNm29pN=8[dd L2ZWr{˻Bq~}tO8Œ8i`n{ӠH{dNpHˡ8{Cw.Ido!( ؼ%i8R$;{z|dJt^S& 9e&HP36ε/b]sb>jUfO(`Liф8 (*|ScBI"E/DMHDU ۹/YGtGsSy8.Qh bhWZ`SZ"q-[M/IṕnU. S^3fGn\sKyNaIAf;EH y8:#T@R i.Uw1#֏5B^g)1;hSPs)`uRv5^ͣ1QB0Bn = Ѣf<(`n4o88'-Hoc0jH>-,Ź"P$]ИK|ȹTݚs B;ڋMEa ͞P$-]YZuٿ6r>mNc<㨋iTY/AtRVK-S`4jm1i˟uBOw~:r̟{_d7޻C6# *E2{H4:HJ<1H,>n=DvߚOQ8;ثe J WVACk$TQlE҅-{ g.'6Ҭ&Ј$f*Xa%M)ơݒ9 ԓILY#'{&>jw#1F_;9#^19u<}(, (Ȣ^AK1JtY+K^iֽE*{#Xp43篇vs4t6J7 vu0~뿲&$Io_DlsŨX" o6ߎ?2E3N@ S;悰V+q\ A}vvڄ>s~snxn80"%g)JU/-*(~gf_}޲8}r ljUYqBb&<=X?0rxQ+}ݕ‚mk]A Ȗ:-"8K{!WN WF/mX2PqLLD'`_. َe:[:|E'-Tbk)ua^#|y"a]kNP2!2%'^" QqZD5`\4CPU#'Izy>g|QhFoXBTIKQ$A@ܧ=|@O^p-tikWkچ^ ؓ{M;W1$\=rǂs X^]U-iQ( tד 9Ixԥ/LtaosgﺜWkֵZ@j-'YզLY6ssg# 1n#bTȯ>D!k{&08>=VrGcW/#_nYpl`&/p1N.Bl51JU-=)lh\ e ' B-QVfRE$1T푢vaĐ 3$Z^KQ: W_[5P@J̩l%0hTM.hy=ci31L]XF]C1py;;:L[U_{XZt'ZTz)5).@<< Q#@-fJs֦>WfН>s"LBbɫWCixe.DjEϬ2gR$U"nsb x,lHǚh ?٣j1x`n;/UM>wHP6/̠1VIޕȏ.҅;L hjD J &U ~:j"J٥z|d)=RnGk䙅R3/g.fn. 9%!gGP8B{CZjɦI?S*6QUOБ7_4{_쀡.qĜG?t=/Nt zqK{c1Z1XjMt®9%y !Z*j D'"i: ;o.dx}=G4#kH Itq{. bxܪBZJT{E7c]uC%V7h3K  m #)1܄5[J:"Mr/xؐ˥b+YG%(dnpI?j'gƺrޣ,wzHv4]ϴlJ~gArr! VMySWOIʽ@]Pv&ɔ@cMd)+?f(`pHyòG ݪ6۞(k8v\wpŬѿ'd?feڛ۶Wxf{]?~hodms=Ix@yגRRo'Ev$-ڞIlY$98/^^dᕋ3󻫯z%$6C[3]KBIĎ>pݽ'h,sp^bZ^d; A8#àjJE"1G·_-QR L\)¢I2KQIߋ?l8/&r)O,I5 ~;zĴ==y݆ZO`>Q{稕md|+LQWh-Q_%mӢ1aTIp,Q Xˁ9F&QCnt!O~0`(&~ h2}wrHNhEbbIrahbٔ9O.M;R58JS} t|h4ޠ7z3cxz ^ahAaޯ=` c/h%?I.G?wBK?ONx}huo Pp9][)48lmR]aV9Q8-*EQ磣|vj(3˨A%u;G|zW}S{}b2:(=9% ʣ`?D|O9B_M cKdb UCE DZ^pWo=-7\QћA\<0aKŨdl~'37-߃{xWødWŮEEϢADϸ{qftOhQ7,5kKNsx3;ݪ_.o\ o\6$Јl:(tG-I}IRUFdLT%t]7j[m!8\h !Vǟ 3;! . "CBLߣXm&~Rn KB*F/Kj!ecPa {q05bI!U'ju.Yei H~ b~\>n`!v([)40"[iVp+ΠTv1ǘܡA;+ KuiLBAN58b 5V@)}I|𡔥ؤώǡڨ@bl/+}Oş`a e]i_,h{ A.7*:^1VnPPo_+|w A4jMa/ Ky{W~گQ99IG,ՈI(^ Tx.?7M4 USL~O^W5{YZcάt3QgXKn68Aˌu`4IҪoҽχwwݭ3ׇΓvyf2V SfRYVnBw$22+ϨC"Rr+5E/]6cE=A ;u)()Q9V9ew \ 3 >jF><EcGSvݷȝk-?)&)"R9?r[ o{׭Jo;5y?:2W-0t-s|PfA`V <4hxÚ-[y;k4%#k-ݟf\NHXj\!;3)^-2 ٦0kλϡ`Wuk#e2i;=-fNAӬ9g͠3awAc1B"V91qu7|7[{Ip- n!G}޾vVޗ~~yg s?:Bڥλ'c )[>fSew7wʿ{2۲?K#Dϱ.-KI!@ڭ3(:o=ӬUj'lw =?4dhzޢ,wD??x5d]Hm~m}/{q_y+yTi, mO菻qBmOӿ ؉i PZqR*$p;%ԮC): 8i^cncUB`o:c\o~C8> ~JATS~n՟()·^yUʲ~8z/J3RB[xBq2B29Aq&QF#iGC{QIلm;jx d օQ$Yzi5ke]bȖtXeW5_lʸM 0#Bbb3q 4‚#e&f9xmoQ+oe\c$ :wkD}Ut51-s8AqeA@p^4u&XZ#2oQ+AoŸ۞k*ˣIΌU(4+*"RbB]yk!>R1iD%sBx%Z vQزve-$C3;ytS2q eO '9 jkFhBBZ[!pX[2vJ2v~0$C4-՚hBqD"BSO|= d!wQL1a1LQK|WQ ?J9;`>c/K_^%=S?g> ?U~h ŏ=m?$NO~E oW{L.z)y?V/N߯J` 8w) ݀ `#KsDR0H!R"0[hU~oG+E-ȱ}$ 7w@G\qud*d6s'KcNjkÙ THfȕOW߈Za5wOo`xv *Cև^O.@N88Cؑ̄;/$P/ɮ>#>|:뫾)^<5j ڱ_~#Uh؎Q @,RD(Psh:Zb6vVupB}NsӎpUhY){#7Qm"x6,: _kxpR/~ "G%J)#m<)ŚyIj%.Σ 'd9@6sZ&c#@ofF".JJW8CBږ6N=|":)42"ǂҜ"QxTaψ( PГ[a9 }F3 @*5Gp @*O4ŜfJ3,k vˋN~g{ wh+!ʹŐǥB{N8[h:Nceff^{jRe) I4睚] tZ vd?YX} }*}Gx#F߭, XwapġzĩFnJ A#%1p X" .J6ݒ#nŢnkV8 FmT!F%m7mnV,6Ej2҄K ᬀzC2P B68';*P[mL`动EP᭪%jVwQ7%6TJJ2\7L.'`\f~_I= y^A@ɸo6}Dh$4WK6.|oN6گ/2eN[CV .coA8N%`֥Xq0GQ!LXEs0K0 7s[mTj> oqS`¥#:`OSXʹ+{JfA핖aO?•П|ONp+cn28ɘQ *h.rSy~%tG]{7q8 1SS/i[U?.Az  @ \C B F߅ -1"`3J.m &GJ\:h_PĊŭuқӍ׵J'LS,:$Ŧ4 ips˔0A8> mȗ񤋎f;w)bR֎m>r=-8-&}h}?@ z7zRm' $09{~qlM |2닽Ol }I:;kkڡ5_pnh֜ h#mm ~E 9k3[m` ѲX<ھk2H‡aMTӍM[ͥ檫| L#O+~Glc@g3kxV4UV'&S?iюU @f *.``RjҤa n'A*ê4(q˃Jpq:՗vpYE4D qr"?Ey5vzdCz`JC6lŲ t LAi ?V]ϴ;^-"w d4fdxnj3Y\)#kN738L567ƊqPBcR+Q C†TMlYn+)j@QZN4E&)̓|q{kgEWׯ(l<&?,16m.&~wq DwAwrZvpi=~|}U> ,#֒eaS:^F0vcIԙ尮X)\ts-{bv;o.W};<QFB Yrc)Bwj9w CJcՌ&9\jvN[4TӸvec2*4:u6^7#B6|Z#ZU94 0˯2RIQ>O3|G00O5+v45]tM. :aI<+?$>a77OxjMC]هSBx7Jbf-wW;GM&6 5K|}C"Jث9:qA0mj؜////aZ CxK*/byS2 |,o~oLmMyQ7zkd-ۏۅ落v [m[-ZdX_U6 c>`o1{;~_㽿VE I.\\\Nҽ) Dq*wWr={dzQ;"G/>&uZc^:O~\zt&QQo"C~Y Ga*[?5:S3*0g VcUJZ>|=8<҆4ъX&SHEmS]ShBz~:qYYG%kJuk pNuR*wWwm7(W ҃Ҷ |wse ],$9h@kjm*g;ckc@^(^27R{bp(|]$60F9Z>mw=ND"waI8h@jw:SÍTBӴ\: 'hA9 ^ÆL sP5,|Z4ІNTںS +fK析ƣYh+gq]۾/s؁n$za=-~J8[C}IH!W{\y8%僗[Cp޻rs5n,xG֊`[#biCjotk%>ʂK bo } )vK\- j9Z(rM/k '-o!qG/- z8Fj/Qz_>Y ]m]4Lޗ(ٳdJ WqHbx$T=Lh:J/ӎfnyYK ֔!ңUR+>lLR g4*$)YB((hMPa.ܮ]ĹKug1ba9w\ɉ+)u ߔm4)9%LZu[umoVPPDu1"-֓|Z-p9Z.9D. 1V[@Dp_mg?+w$$+g$ bc P~E7Ft.M0A`Je\ssJk_iBHi0F\Fi 7NzlPެC I`'$<Цp5P*$}75I6&=&wc; Ø0~)&xUäź>5Q|~pLOHBTLOXsѐntK L7ރiEZQژͫ0eCc*|9')R:(:IοbÇ+ b [yuZ, ?_8a,//<'~'Uß| ~ ?Fއ03Gctr;]Zy-2|:Ӽxc3%HR2D_cP qN5\}A~@Z#arS w>n(!nmڵ+m|0BKlop L>,5`|;ぁ6h Olb)j6:tKBN8FP ҏFڶl+, 'dEw+D>A9cG/V7//hMSί16,21`kbg 6,7?JJ9$ Ccb6p!Hwe X)CFqzX5$!yN5c$6E9xhNr:`'ņ4|5F|>[eCbQu;;ɓiVA8Ow>q;CZQ"fmQXmo2}v!n|HSNbcB+,̜X-yIKKZ•d-(wEղzt[$4aR|2~(8cv6wx"t7D?եfv~pK&R5^x&U8t]mQquQ'Ç7.!,6).$v oFUYA6wzT AژOT>ە4kˠhyz<r^4/B6[}rK-oG_N'.;4͒O<~N!} EIFig4%S(cq8[Q; 58pKͺ4$rfWQƬW#54{-Ǐ ;6,77pŸf>9\fEW43(eu?˧W+M`"k."һJ5!pwUPI0RrVW. ~8*>F og ެкz/R?''U>=WU/+MnWUpRRϨKIzfbuGM_㏢.\}L-ψGXy8B2cևBv`;_Mӻ|@.cʅg~cCge\g:L,` ꙣa@TLjKeJ-pJk\G u? 1ꇨryv?cx[%.G˭d)!ft-K3V!w j * #zLȅd5q~&u:NuSYHm @!3yz`L/m(;71 HyfT d=W!#SjR,#fQH-\ 4NSG)) PeH7*yj3g X#TɍhVb)&]2!RHWDq220Aq#66;Z6^ 4fcƘ͘cf"y.eYft$hZ"W2~-MQ=˹7LB!yvnCd V PG 2\BC򛐼Z4ѪXeZeɔJ"RψYT 5HBtj[NЍC iُa~7iX'hmJY*)RI:kPzI㬐9nRW?Q*)MsVRE˸{WƱe 3iE@><$3y 1gZṗs]ѻ9GTSpce_2Ziٟ|8&XfbBVCID-Xʂ_*DO*Rbr[.L$ { Xvnըu1ɔS,xB**3tP>qDƌ6Z(b&A 0jrƫ@   6TyE0EqEx&P+/:WȔcU60<ۡ5 (_x@>0p-S!MK28@;j80QT,Y!V&ď9$Eˣ*[IoRSQA?mh(ZIVџ1T1yYJJ"f(JƬA5̹Ȝp.EH.9msEΡd0(lh B "!mXc'-`u:U%*\!cb*!!ML((py0I`PP*fR_)_)e̠|Ac醳b2K丌^@c|Dj3t+)f 1d a!B1cTRyD&XWMU(`GyR5Fa@Ab"Q c9_R[RI(oJ vSȆj\ʀA[**dX)@{d*3j(Q.0QQB gHQ_Q x\*BZbPW&[`R"QYHIBR%eL`#/=kjEQ_U0Vr)F@BI1&Z*+T햡 /\m;Z댉0Ld3& imn{1".Yf͘UQ$8_S)*)_,Ka8q!1ȾӢo{7` W6k-+Qo]:nSGHG2^ RX8p`svrmV28Ut)\]Hh)7UASc2$3(v5K :".Q%*|"D&jZV"aA>X\Lt~l;m<{ ċ ":t`2GMۀ8ڂn+0fXtX?`<9Q*5ks8n𠗨ś !u>fTs^ 7N(0j@=-`u[66I#+dO *o*ϕ+"@r iQ'yMv0v*B&4HbQ:TCW=k@Ule?;+>aE`Kl5hO fINo|ըuȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:O֨!e N.ެ. GC{hedRr,=<`V-9W@1}+ECJW,=th!z+zJ1JWar0(]RîWKw( h| (qR_T*hd|iO!-=:}8/~>~zx?ycY3ݛtOGЮG$7PWPfm?ɬ O=6;2gVq[/3~cVBfyPcy=R蕵{ ɯz׌+.3nI>_[] ._y2~hWB)SIUPR}E|9}3GZ~?+}]-ǶЖ\7B1.Ҳ0nN,05N|d֒| Ӿy<+ jM{@`z@Y;RZ+RCi U"ɶHPj{D^_hz vj6oo:boͪ^ehֆM?.r눜蜎Jդw͍fXimb#e[]uӓYU0rޔU){'fgfg2s\|!2I2dnzr۬|?t[xmlc-_!|3RLs=`ߙd*Ytd|:nh HzXt>Z39w&~i` ͠L30 1sϹ̖o?ֻm Z^7ύsC>Ƙ!J?.|g|qxmNg01bŅ.ÆVr}J_?t //9/Uz1lgɟ?w7/~֏K|62)rzi,tvr6&'O9kG헯Ғߌ"hw^tK_okߋsfv tPs@՜ʧ)p}<IJ[CoC۲l{] ՟mOV\SeLs+lrͿYټ-T_ߛWwo"w_6^j:{%ymgU/]roVz]0Ԛ'7?,fg/av{+\jrSKlyr$vS x>ӣӳѲ4ˣ !yv.}lh\Hf]>Nuy?p-4mר6o?clqѨ؅۷ֱTd1Er?\6BC_,_gkѵ9R䏳8LqD_.TrF}KlRJfӚ&MN١))zTM1)bl49EShr&hrf<=z@zնcop҉nJ,CC=H N`#Umͣ_ V}vf#v9Î$ ~NL[mD|h 3luP G8'Xbc&<1o] t\x[gt"-.yhF&s>io01%l-U,;R+^g hҀԤE6@u\9`&XI&yr鼾Wz\;7"AB&%R3b)5hX0Ԩej_)l ? P!~ZEu+c#}x4 !'j5xNČNj0H=1O2 ER3Y&W^P"U[矿kl 6n%My%^*8菊l?NWe{ɫ=c2s??cգӅ:'QZlBRD;EpI {pHV:ʴjQOW:NWQ~t4f\7]4qڱz!,OO'fX<—77mށ3=^Dd,mvFte뻭pi8WfI$GYH\Bf  ffӯh_ލcѳue0}-0n}ۼ=4jft2D'lů| u>䵣#R;M4%BqƞDz\Lʣ'`jV_־YxMb tgD8&j/9G`l qcmd> 6foyD['L)_\!,*@E r>NOͶBQ'A# A2&J\ E3lHf٬}ПXG1aR[Ԋ-NF˙bY;Ccr qcYΧ[[tƩ9 NπkQF6i/+)L*Rzr*fX8\!% Ќ% 4,L 0>LP="'$7)!EKJ JuwZ̸kq#YtZ~@F3aג\A\:`ަ-9g3ft8lӱy%qqvh7s4W!S`$$%Ka"VE#0uU4 |,͇%ć6EIrH-U^U zS,eh3  |,ˏ^S.7pN jwW|1'zF=4j O>>=Y.]%F!rbu̴%ª#=4yDz|By\qt]j#]-af#lM%#,*\rg"B̰@U n2u ` ,hl 'k=3MrCcr q#Yn:>nRBSvӪ TyF$b6%cL$!+uscCoJI6ӁtDz\iC!DsQظ" "*$sM+GںA]>v:W +} ];^s]"< Tq"%J0i)/D&y{ncPtmlY P,`<8fU0$-fwh3GW9Dz\`_ijU>C:72II% JYAC&''Ge1c٬]53T6&k5<5I Xq>cu ݎ'I C!M}~XU?-mD4+N٩n7 4IA K p`ǬW9]B)E (v?Õ+OW\s FE2aᔏUZQ;F_V1rXF=4*sq@M)FMxzR<SmD0&Jzsqc+r(AԦN`y\zl!fhW[rwF^쀓B)`'489D ѲrX;QM%|5&/f3w/̠_TN馸\ B@hk(.fi}a1 I!uy(L5fڕ{ڂΊK w3'0之5U{E1aAP۔Zin)inL!NirKfQN/fђŀeWd5s.@ %L F6QKNٌa3|,'M$9C*J x&dg'eM' ch[1BGVI [yyzMs+䢾(4|oJCp+74;JIftבCe7aCj3REzՖ{E 'm[-B孙y뒣=Ly+ia>54ms6E_ ~Ρ=먎br_jYifY}Ae'UX8r\CP~Ø"Ԝǰ67>%u.)Q,aO|WNȹk?!n c(4HEZs7%JydiMd.q%I=18:,ǁdӱ<@"<~+t>ϗ\CU̜Waϧ[k=4*~ZBU)|ظa~&0,@fl5}Zhڊ9 |x=Шf6iQI'L5SK1N{G&4UaTG%GS^MՂyƮžOXǵ{u_rS5;|˂멏|dy< ?Mس*t5gTA]_4v6!9F[78دuβV7PҞ7jL n ThM,To7epb98,6 1#>Η#5Q> (@u~Qq#h-ttNgW)T L]/U/}-'%9|> 6f1jg|/ˀQ=zhTK /P5fХͪ^,Tcc{͊]I[I_ʙ ;h5n;aÒb ";%Q{dZM( HDSb8gHƞG>n!un4$m!$p`wADKi4R%A$(,qu8;I,F)o4GSQZTsh^h 7v@*l=uD[t$!B%Rԉh8x(,U)d& 2='RH\%tO2fh]F宝y=U%j#d;՘'W C@d?eN*i7)C+?0G=4jdXhPy֗L&Ÿ́^$o:AR)I]%V&6'I G2>j[7xdx Y#`{}`Z㭸CluQ3:QqrߐN{G|Nx1 ǃZBGHaS[ByFk8f>Nڄwu.꽋nCFFtnLc^mò㖊;bdջGXY9sz.b||.ǭ^lqc>tz/yz>Wf-?.V[s|m1"/lL\+#1[xD-ffذfs:@Dۊ ':/L+aermRNbr9j +g`{Cp0d>=Zzexz'J V(>QHsor^^ z'Q V(`DxjgB$Jvo"Q܌&⛈hngOm$:I^~  Of;/ k Szzmg[O-Hkڸ"+n+VpX2{Liщ$znw"kE4d2G4C鷑-Dk&_1^3ْ~ Kiykri Etg0[9+tg,4\Hdr s8t o#Efra+c0[ѡDwL~~DNBUYFO'T`R"0 u+>{ȼ`u / iNh1~#Z+>WfG@oGWlJ7gʨ~qPu7dtif+.m#aWYΜD#h^U( 'dߦ |ezd6!]{aNC~H ;s?rjX8(1xWsTΨ0QMcZ;_w"kEіHH$tYr)O]gWH@ϷH@;g<@PӴ.<0prrAiE)q-?򞦹uS=l6{!MrzeO Y%VS4Aƞf7 4ys-}G{ &2y!˙7VMhF|!؏'xG0kȥywO#l^(GP\5]۴7!:@x'y]ζ25pQNd$g׭|[Cuzr/0oq$sMa|dc5Uz9}B[珫\E0QXԾx6P; kZE~A"gf0R|hgc;/dDs>ux۞nnH 6vrlm濦i<'$9t ;;s#bqʠ4 ! L +#„4ed<)ZIFKђOVBQ)5<*{cSe }-o~Z?1~;_W:Y06w R {0Ǽ䄢?W@?!iEPyl.('6/cbdL4eB\j$Ua+dry"!ЂAKy1kzDG nHTpuAJAHڦ\NM^.xu{&!|]hI1yQ0ϲOv̀¼=^iCȯfyʮSvn)M{ ?r'2ʏ"P"CK T2L^H nʔ N2qzMz39{@g; <P x"`,?$\c1jx{n7wG;:Α dx_&SMeo\ePiHSSM| TX͔TBcra}YΆcyġ]4Ү1Kǐk<i@fsux_+MkE)lTid9UbȞƞ5{\쏶rmwfwZ+~N1Yi~Y}!x}R-ʅ[QL-5 TNrNcra}Y.秝v0~|d5&2Ymå hg9G0~,V, gм=҃ nr25~~\e#o ho`S2mr8\+ޗJ"uhsI(Ԗ / $k66뺍 f-tD%w3\k(#DGnpr޳m0,&t@t_ rTe3+;fji1a%怟^+Zݼ:G~\ d]ȓ 5RPZ<oE<-Lw# OC?"^4ӝtE$PfJw%/Yvx0;yAQ d KH7N?@}Y>zf A3vRsqNN8zC 2bn֖(r Τ) F/HޗEЋ\'n-4A}Y>x ZTS.|ms|Niz)XO)*)dठmz/wʦo/bPub Cke*d="|>jJZjSqMdFXAw8G~>_@}L.LC/8dx(T8V qe!Ը |/&:|׿ h:Ik9dWO |Tc* /S SA9S'hYrT]6D'->HD8^ڠURM0Q idczΓ1NjIŐ}=ƭOZmwe6ӫq ɇӛrBabiG[l:@Ƅw9+2ʷ! 4h:dmꯆϥ8\ޗwOo57O\]j Z^c|r hi$b&TE G$>3hMU|\c2 /t Jtlz-@gZ[A+xO6s2J(m֜)>댳M ֫}SqM NIM؉ɸ4.ȍG? 0"%}VS e#Cc=W:Su^>lˁ˨^КtR5{iE]&o-S6ŀU1?:%5pV^| tU2`.fS~ˡ5y-X % b0ٰw [hLTJ<)Q, _/e HRY},*UgƞjpeN]|rUu ޖ Cy9_R'k#L>M`Z,dEzXKeAN%ck G-4{@uT^si"]ʹ[6m~';M_>TR]:~YE z+  5/sYmoռo> _? K2ZEZ|4J?z~y֩OUe*?}p)= ׄ~2=L#5W4C]Gٜ%swdw!#2C'uhGquZ#>>i;)nYFw僸;}\I OV܁6 x@Tm; amÖiTGJd꒣-&*fqj&G.)0woCGv$30:3oWwҡHd9KX@.t sz97-! ͌ JۘZd~b$E5 M\wh7Ob*C\Nxx,{Ǎ/Ip~?X· Yy4,qvvp=ݤ(qDQ*a]Ū~TUC#cEJXmucdp=֫i<2k>lY䰠#7 ksG⃯$ibX g<]W=N!!l}3"Sj#sA [NFt9-1I\PbL z/Z W3ݮh 2T$h2t,ޤCZXߪ^4. $ϛ( t=@5Gt2bn"2zY-3\I=Fpo:l:6*pG(d?o$V)P07rHKE"K?5H[I<3')F#aA"6P5O#s~`YA/Ch^-$e8l9[ ě랹/OT O(H(McБ9HX \ih\S]5]@22>G$ EіMmsDGePP#3G:}iGPQϨ: ]bJ+j:*I4O_z$)B:p̜o[vp?сcdოw|%hi3.ЭCI 7ed^Yu) Zv W4I_Hiy;vlcd YnEyT,E)jg+%/)ɔUG2k9ٛ>-^#3o|(/G^Hsn9Re#C %12s*7ii12s$"fƴ/ʏEɭUh>ȬĉYLëxp*XiX k(ՉN%55p[h^T/lxL%c4(Qdqv)qzq3(P$nd#Tp}u@g;p>@] K%i OQ̚$|ؑY 5P5f-߃Z‘THq: Q83Lp\_\)Jfl #&E52oe֡<#ֆʦ&M٘mJ<#f8QN˶vcd5' ౎ATO!yĤK 1&N渮<~MPES/A?#93[ 3#!L$MdU d$!C,V!xI%FF]9r.c EP#nP$( ⠋nHLy'w=@N#3gXeC ITb;͟kxث1Ϣ5tx\U12sH{A=PViQ d:xdNۤGR(Deҡ ,SD +iu`8FfUN=yv9=Srףۇ#DPy1J wIJK-^u:c(T8Feہ9^eJسñ⛟2f=Y]]pѺwl[q% Q뎋0y3n:Efoo5Evnaa$e[*vڐ{A$\K"W:j/E)}lyPx{,(7q-(GAA;LHBjG8N\NDy2.I%)aJPǽCNy{A= %##RA EzmB7-qIUSLMAy%լ>enY/"O+'Xj!r~ ,+"z{:%mu 8;O$BXF*N@g(<u4lt`&&eqTFXʝKr~ُvn<6 iyZGAB;Ͻt\1 Yft'B1:?TdVYT7I4gu^:S,J,wMnk"9fAOgBF~)#Dm=.!QO6TES[XxߞR}i#Z Sk,A `Z1j$:2B'/k"q-Bk"chh )P=G而vw.I Jt,sOކVTKv;W Iq D*2W)q޼6#GBʝv^=]*$\ki dLgRJ-MH̋7rbuPZzAZҜ`?YB#=Dq'8xE ]g3!$I >{5~3UO `D4OaEPAcŔj(}HDpg 7R)+Kujip4tų^'%co[B[ADIB׏tO.+ "/iOKl"l/z7[pd/;_VbVnζYz'sx!3_g㇝nEcVly| Ջs -~K]"*&?d3]vWj i™,Wϋ/MݢfK1֠4vv)rmSW &јƿ4TnH?0 :zdL`G((Px ;1u&sI]^h#j,9uB ֭w0G;ƣBa%HTG\K֍=)'cCcHiG+15'7ʡ@wYi+辽dL9-P+_ٙho 6+ڪvo;RZt0 !fHgz&vruK|pfW.'|e)x7Ítm5z]EAm 7锢g?u_.W?*tQB$WgLS&BWdă@0Np"X+;d~BSk'F""h!gE8:` Rk-\ E"XG=6\;ސ7!rVDՔ`#o5y/B2!0'LeK^̩u|m uTdLjC +㯽b՞̒ {25(oc+ wcܳ/mȫ$(R&j*KJU-b_aU1!}c*r^˾2}!sQk&Qma .N5hwf9/o\YOq).rtn \ \4Ԓ^6<@/~O8S߆pguKE1 .] XQ2M ˱Vm菎K{0T2}j)x<.;#np73{ƥ~2eov3"os. .*(.*Э^o Ȍb[*S`stib D1T=֯/D#~) U7RSMp{~rʹ~݉/UyުN^3+1S N1R*(hNB EK{@i狝&<<)TȆa,.%W)MyC`_yrth & E ]ܛgІ2ˠ޸92&pm#c} JvL 뇻L] S_u<5ꛏec6Eck-IA0؎" x ~z_sKm0O[$A׷*RgaNo{dx\c{VO}3<9tHL8|x|5:9>&T91dK,5%x ;cY|D}PHfH~a,ҮkL!(9Jג Nw22G2G,ZCty@==aM9eKWg3bT(.R[`G#)agXO_py`ۡcؿY8K =c^W+Kudwkz^[l yzoR.)so:kGEC:k]¯`M9f͊0eDJ-Se:)IMt Ŵ ‘9JW9V|@kFGЁ_Ĉ\ V慟wqÝf^ݞâ)R(CpAԲT_..an?)`,<.#³Eq@×ޠTh%D#7<^W-JNke aYjؐxKcpX!+dP }dL DY\$evJ!*4]Ә^B~$ads R YuU|ojMlр5uI@-LsFNnm=}q,_0nGʔe1BO>!q~餏Sx+q҅LP?U 2nbm7 g*8A *"\B%nj_G s #cÉ:) X-7eԪ5Vҁ79л7M }LaQ3>&ey״2QpI`0<³[H WQPi:W2Oug^jm`l7N"Y|CրN\Q<_ܱzdi K%@` F"P*GF[׌7ԛECIL:T61 HƖÁ (Z܀B鸖 24`VCѿ ҂w_hMJuLR,l.8l 좖1؀ v-mU50rr9l@tI$ԖC&r<'Ft=!  \J$o>?֗}"H "kea^k=N|JX-|ňgqk5Qxo5&fGӚ `1`.*N_$ց}e0"Nג{,]fjv\/ѹ)͙]PtvtzZp{YVeJJB5kI4&J~@k ~iÐo 1V.RQmұWOfG"i>fVE2~'zn_y@~xhkU7Zg5J@l$N ?gU)FKQ0h}GdN*vAB9+ wExJ|SyFu89љFyt~~[Nntvʯ7aT'A/;QR;).c&eM8~n|\gC&Ӟџ{,ѵ#pYY%/7q# j$Q&A̦RBR3U`X"4  :?uF@= ԁGo[.<)z5ŷ&FtEpy,]07p+S`s)^* D1r^arS,0Wiܼ^A+vagWL^o0Q]ؾ@P- &#pW"2=_LaBʘ.p;k Nd PUG9S%f}E9{eF{ft$Á)ȏ*La f=QlƦHW ^+O=btlwFnF|ݸtKf\nԎmgЏOUjwB ,q)uh"@h딯q_;I S%Y Q/*pӁGS%riС`aJ# 7Á37ؐp3 Ʒ^idDdO5ZSfp#mIds{C! w dܭO5E$X^Q8)>uWWWu׻nM/! ,0ĜN\2Y-[Uo!D#-CʤkㅑJ J?vCPs(c+r+kuggIYqx8`WH 1oIGŻKVQ?{&W?b$HVaj1~Lӣf?+@5#s.G\t*x)Ө4J`@ Q`bM  O]c*5( z>e*_kYǼ8ye9p8Eh#k Ji1pD7BD&UNbp#m6a]'%9* pfS8 |jUDj͔#EDM7{,|#O櫊0鯜L* HEQ20|eD!NI-#{8X*pƇ*l^5d97I 1o~S?IC]zGo?oеuV?` F:-u~89gP]HeA\e Wڠ1BvK?|.6.|և+[+ cX! C\k.$r }_C;O\kA :.󾸑e2LmM}fo>5fffLպYl(.jeZʘ^8G{_< Ǹӊ+(Kt2%)1c BhZsGgPhc:WtG% >C)-s&L2'W"C0 ZΟ؏A{0zf[XlS4`.d2x9qAlR9^)FR )5&HU`(PG ODJNcRYO@bX .H9 (O 8+cvLd0}=7!B 6IY!vZEHiA5Y z$> msˡEl[03iUu %IH%JHݘ%&[5c$N. ߞ 2WɊu ˠ'k|ZNMº34=IӜ`c(fְiB)'2.Uk[E]DL9Iv @G#L]{)(d.7\@gJ14ܩ`aK\:$UX u \xYYDxO#,uX:N>TGYFmj>==^\)^21fMѓ`sނ)[=.f;1>U0W1;k]ԯfV_'?T^_M?>8k+Ysa>q-շmhմM-;BfxLb4d4A4-0Qu>7n戞ٿOXj=<|f]V+[:I>041hWPbŨvcAC q?I߾>_89&/gox 7U[W]׷unqV+]C}RGk..SM8'ErIg0}h ?Xܟn6ZK{@-7~n?B]M ͧu+\y<(q:ByWլ^#僪tF]2&vq:3k9A K16zxݝHVi뙡E@| -T`[˙3fY5_@&PzgOHB?%aFm8(68ec2@!" ItDBgL'^I^]nI ݭs=5AtD]"HcWQ,̍FAH*!X%3w̝>s3w̝>.i(-s ܂1`-s X.`-3=\n[0c܂1`|~kBJ#[0cn[0cn[0h"T؄0OAu*oe,:u; y LqˀY,C2NcʠXgtcwLLr.PHx3jg>OLj9+j$>u4p^@C;$a^]Pfhk($Zc1tDՍMuT,cn%Ы%(ՍF1vW]z@_ 3-AT% S ߂>/C_ @1'[ucO8 yS8s4^JJd 1Y,-;Jaظlv8') <(pC;VfUT_Ϋǿž}  J}/gwȑy<[Ie9YW6gQ;b+dl9ZiDۺu3J/&fbLcCݕ:k⊟gMGðj貅nJ0f~.c.i_6]\ <_ջm@g'_RfP>3NJⶁ!}CyvRV;Yܛp1ySS͊$U^ b}@ #_ y(S[s4i,jޤ{vu&Mp.,-\U!R<@<z P Q8 )&iTPaUqTQ3K5q19I#Pla!*[VSӢj,Z$[RY앀JfZYVvz-Ȝ3ŒnD%X %AIv|.$nc=a*C6 8%LTvDq{o+uE1CYzu |1h*p~6RbǣѴtokJhgLجWĄgA5H3.//[-)[锼%A;D O܊ !Z']ИOݧO.$jr3qAid+Q!>礹}gff|9e\D.ntysfhP!9>$LJCr|HfCr|H}|Russw99;xw%і!A\{DdPJS/]q%.^\`dkx.Y(2]rQx?<qi_ R.,AE\N- 'Ջ6M,}NZGbpKJVWcq| x$ )[8ͭwZ D佖DMFSE*p.8p'x$Yӳ] gHjXF|0arroDuAk&_7:]͵>ܘvvZچ<ن5 dcx)wJF(bA3brK%bdF/cL.^吹%w- Jk^%^lnQIN#92urE#h]UmuC4Dz+j5Gyc =^R=ڇWw3*b9dHT-8j6+kZOQS*N;M]Ġ9VXDş|zV\|g"zP5K\_`OO;+_?c8iq4%eˌ[I= ^OYOô8{ci0.RZiUivj+j'\1@093b-*IwY\4E'ʝ^{5:_VKfjSSPgE 8X:+8gE#6L\uL.rAT͙1ffl%ivwN̊MZbSxr]. `_E1d2 $}7C,Ƿe`#sGJ㓻u7m ~{TTAe /HZI߻+MN^[sW~"|; mk#7CpK8.X+F0ZSmųF%7ӓ-P.9f CΙ`2#P#YlCuI6 Nb}Z)R!);"U 9f$Q>Q]UUw)!-v{A:Rbhw)@8F.ٯF^jF>o#aVW$Ȼr0WݝT9u 6 q drZׅs,{SEX'LR&5˲] 6U4Nd΃xs!CSω `c2SNHRjkLDPdd tAFW ^ *sT D@TD2'@bXTq`4H́s.NHgciSB’jatHP=\!5Y _݆mBrm0.αx?&Ǭ>\gPeŲXh<3?/.hXSt0N]e7iq^i <]kAɶx;+_>ZX>x|MxX1.0] >yI[״8ƓwWK`qGa!Tǵ=%.! >fY^0 (d<^zsx1ᱲWd]usU^٪Yʃ2:>L]jjsR|I(n65*'~.Ǚ^;~_}*˷o~[Lۗ{7_¾P`^͋ [gVo~r\UrY`̲t6AYf~nβeZ{d@//kuF߲kXY^I)~kU.|yCLޤJヵI~tqXSWl% MtL艖VG;í83C| -T`yh+g)yaȞ@!ƽ'VSfkcS61"%BL+20'f[6:jh3]Z8mޝ*nfQ|Qk^ǮTT3T *?ZlAΕ-]rc CT6J! ƙARL ۫Qԇ҃.= |l:4hQu ]FP>0t>RCEa^iGش%#?g)Oc.*5X^1V[_[[=j_..O![Q w5kCNٓPFY}꣙],vIwwۺ/XYvv-Sټ;Wx2NNLm6Gwܴ,_T}bVS"V̪`BUC5( /iugzf1?oá[rÜ9Q [cI"mC=|DNjژ139&QmGs Ȣ`;X+\$JCF{]F>O Ngg |3`+;uDHzqX5"4׷O`LW3صdK-U2 '/NʈʋoUC 72t5'4{ W釤~uRX]{#=.F&^`s)Dy OXNJHvFM< nøwm OIYXx޾)9ͧTFyeߗ~QYE(B-fengn<kqn:j]N3zr5] HޑyHoTml٠P5b.Y.`&JDQ3tCt 0q(fb]ZQBi\[RtEO1ohH6ߞYyk\~Y߱R&OMћ՘ObT.%^`B5,%a'@ÍK` ;F$㑅H>HԔ1тFQ4`pH ܸCgr:U{TQgۣҲDfKXịUh4BkM̨3hQ9b0XJBԀDL,^cnCVώ?qޚFTe3)$iG2ױA-$3sȭ"G*("$>wfk5(FX>^JXyd@f!ws`dLoVa !՘V&o6^0Lc1>9 E N&x[Nlm8Q2r2맙A~&zb\ΑF 2@} #.:SWj1).[+ep2>7!H/)OGRNI#:&Fs0aljaoL"VjP G$w SAq$\@K;ƥ XEL!ג)E 09<j'$#$F:#{5Z1v*@6aV#JăT?y^6&LT֏R3E9AZaL0yebշ?gW{APgf`"h4=2L@ap-W#x<(aF03BQl9;g.H\vRR~>? AHRVa S띱Vc&ye4zl5BZ"Z D)|A(8ͪ[s ȷ˼>›̖KY T7E{h3AV{NܼhVKųƃ;Ľh~ia| mYj0 g1sxz;/sc+u^ā:L3=n+\NjSwfhb绍E.*ȸF!7a#rovI𱐙uZ-E9bΐJ YO_ϗ r=77!Fd;dxt]eQڥxMGZwDjwɯEfįXqQkei75y^oWTwq4i18mLw`qh#oA3D- Ʋم9׃%_r8+=^%je6\_\fm2d2ٜ,ndp͓EwÔz!{M'~ZhuVOk&{U_Nå[^;/_] S?g_N7HhCJ5(򣂞')I^zM7i({V4~}Ma]ybTal`1$GHeE_?acq($?B}~ZrYvS/<6}ꎺduO ѩ*4\9%u@v_r$#w`/=N׉иkfo4OO eЖxwX~٫Ȍ+5–5}"(Sp4Ztf}*)w˝=mxa:5PL p AVҸX_bp;Z{Zpw_`D6k|;bzWh2EV6jǹ0 L[Jlư (:D;g 4<I"C0 ZtuÎ"%6!OLBU1k`8iADX/5^#j+=p[yEG+)7]Z.?.O޾LVA尵Kq:^]쩙n(6y7Yl&,'s}  zN\ќrJmD* /R[c"T"; sI ^r7JY!%HAHitL*"Q 1,*͸@1~*ʓH&A |< @2|:rڄBlNY%8BbZlT a,DEKFسz顶&!Ω`,2j sTF筵Zh`wn=;<^\J1,{8N aľ 0>ag[>$. e_yK**983Jgg0 :>[ ;; ޵q$2 `,|r\ b+S$á$ˋ[$ECI&ey=U_UףXaݙt@S7sOC6.Z͛ji *?NW(itY> $RX%E5@ORPk{*F9o%14`jatH\!5o`:<}:Kzұ7ɀi`7KYDL5n\ xRΈƘ6EGݢt7)rai6%9]y>sa8˜Ogٕ?K/՝卷yx.0satzt^-ճ3Ue~QY:o:{ZSu'Wt93Zzr}WYuo&]\tF'+{%hsAuXW6HDFRKM_!,-Ǹ~U߽n775*'~ۙ ΁컓wK??=~ uv?'g޿}Qn\OI(u(=~u ׇ7'AڅRr6;`E)27nb'eR@\7jvKpa (؜^~QQ庘ys\yo1v\.Tov"#wV±1F-gv8s3L$))eK8z$u7OԱJ'vx_.|p! Yp)}袏g+ |;9{S؞ SSOfjg?UʷߔZAV5\'L>nV)r+*\׆^Ne첅6ǭ 1_{Icv:p{Ok+[Z8945fSix-;F)ػϓVn(Ń/`Cyo[z> mf;}2/O2\K85*Cz Y@SJx-Uvqjhf\eWz/Ώ!m".ej(АhC/s ^CirieIIN+lXoks[cIFhiˮMg`]71yrcV2$q{aw,rXt/v5.8bb5u)e׽^,+{A|v6 9n\2 j&: <^\f lw&fz—Ag8=NRP,ٰ? 5cJI6c-X9D;NM> UV&@:gVJU8cDwg0JZ Y눆0EN93X?p2 pw#9M9$ :M(rn6#"( -ְʹHH!A;"eF qAµ7 F;d^v_}"zojVv{q9_Q@煷p*HPU9N:D۠1Bgcl>IHQ E[㰦!d- Dm0y:|lxڻ޽p|ŀ>UNIp|1fD_VQ l +%J?U_F|t0ba:㑅H>HԔ1тFQ4`pHtMܸ4/5`عw{ȳ4[Eqss9:?Xai\,6wh5Z`(3*h2:ڀmT 5 !S\cS.vz`Ekx+m:ht:AJ7W]sڪY9r?'Ox =;ZU>V6CSSNՇ`n]𷽣 6[ Nwlg_8 )U /:E)7a2u`H{2C`*,c"Osr>b&;@*UO@.?QkRG] j:Š \s@Qd^V#M:Hՠemеܟ@f 'BZT};*%Pp_.cGL AHRVa S띱Vc&ye4zl5BZ"j;{Mi[wa7||խ.ߢi-Ub̈́![MwFe)ۚȍ}9.I`hsөϪhńևJ9<]?=WOonW;u,'sxrRvxf7Mﭖg8|}`}{ffcކ]2Y~;r4 ʧ9mnv/dczT)CZ?*/=R6h'C0P>nx ( Dy'rkȽ  )^[1K|-tB<2 _sEjέv`dLoi-HsDZK-87 #$)׍6^ԖiyKuF'N=Đia~9N`.0 ,- T)Jnc;G%}xXח?^(֢tlGm7 ,I{&"k/eshkKZH1LǫInn^n\ubW_w;ŨYzF&h|Oo׫@8q|;L`lo ʫ럶pG6WHCsKFDoAƃlL8`aÍӨc$%T8 yS'Xs4^JJd 1Y,-,xGI4 G3& v?HOoDE{xU2QGRjkN#:Y*j2õZ1㉌L=zAM% s4} %LA)z}ј<@&Ћ\6+k`zi++mрʧ >sD#Ki(Kon6ɟcg;b-gX\asɔcII)N,FDYcmAY?3Woɪ(R^/& >5gnmRlt-]([BŤdSJ1bpq)ɞ4TKbBghG%T(4O;MB8)uQ4Um0\9~@h%;=?oM; l~oェ'XT )m $) ΘJƸP" YX4W_T7˙,psDXnd4٠AeNIr C*3H+]A;B #h fuE{jbyMj%6Euycp .3NB )k(?iWKSd5i RBK@+%뵎)C0к@iV''uekuvv"Inq Zc( >~-23OGl*UX8pN)޶&l$8, ~?J(O_aFWCR88mr6IZ8ܴo̢;Rc3gB0t `:C9=Gϖ?FIRR2 뻫M[įnqř"A0-Z>AZTk.oTYFö]xtt8;=;1&ћed6 cP'/91os__J130-@' jkϾ,>x}<\Xuvl+hkv+^F/%ZRlcKRotS3|c3抜l檰|I`4nSN.=lsx:δֶݭ.AnjuV֫\:-)#a$) yP!;`q?nUw(Ć Ͼ8o޽-~|u?;~1>{A,w(7RjՅƕ.l iQz3ƒY( aY<AZ~.Km1M[ZiloٴOӲ&ŀiԄݿl*sJijT埚+=>h Me?pÕ'?V{Q dJr}B=z|<&օ$1 {J;mA ]{9ˉI_O1@!ځB_pMrGц\cVs'mT]*2S(|j2qG==t~$onm tmAW{3CG_% mICHjKx㑤I5&G:,U-\ b\ zJn!ŋ(MSHe>ʁ b`ˍ ї1 ۀmLEϨ6xl `:Ih4WΎ ;Pj7^9 LTN+h)&X߇B"D/\D<] H֖%H$hU`7_No:de kTd˰,184I(c:9I}9SJBy*)Jܮ#3)Zv W\:/n!iHyO;M#8u!hBjhP'Y&v=(0PJFDqJ+70󓧧Q{v8B^hڜoVSfR32%Wc4eO_b/9|@Ch(VMcQ:4:m'ٗ 2J5O9[4Fu`jgid҅ԭҢy__^_e\'qvwAP*oqma)K-̘2;^[GXS<@(i[Ss]9e{[(;n٩os9׏Ak)HOٳ$\;9Ӌ N& FQA(W5W~l߭Y^I8g AUGg9*3>sS,nq}拾_C N PD cHpђGT>ju0)u qltHaRM}(oK'e1V7NN̅'qħ~pJsHI!NItirK8ևBR[. #Z-yJU1B )[jgWOgV98䖡&%v=JZۣ&C=.=(9-:$MҌ>hrʒ| U[vw7_Pp d|xm/7mZLPRj WZӎAAa1dяpE@P?h\uOh QŦ'&B^i@"֢Ǣm% t\mC xK냷W2A)&$"8iτMxmum4 k4yK3y0;>*рϴ%#X2l( ,J#ЇB V-h1)чgy\EPVʬ Bm< yQ]QGQ ఛhދ]\20tdT>u>W6q' ^q'~aqCQMAKhǓhQ˧1~g6m #l`A{lnq4,9qp1OV]=χtbG?^Xi*ˣ!1E$OŮ0&JeAF6xm،L);dʔzcCx#)Ƈʊ8+ YXLuAxTM`Lq%X .!sr tDP޸ڡ!hCl(reIyeR#p 5+S=LtLp@P9q⃤1m24wECxemj7\H\x"%PWxJ10.3A#iPޚbK1jHIVg*I9)vPZY[G*(e?9GBYZ\]CxW_2=uD[4dҒ3aJgЉ5PVϩ 4s.]Yҁ3bN[tbD;| [\nrđD'mܡ%>0I9UƆ-a0 T U /'Քdoi׈RT2T}(TJ:NCy4ΆqJݖEo3xO;hT_^QNhΣ3s>ȥ4[ []\Ϲ{^/v˼-Z㷓It7꾭JoY,W^l3V?NjzoiOa|*Sfi $˒ʙ#`"s8FXMaim?\߿Bmv5 ?۵g`U`HuqHчx)9ȉLm607n~}rLFx\Z$ !..Rn87钕^ރLzz&m]gf!KmJܟl,it7 YNٵ6X9tN7w-H}&|#v{,՞v7KleZ*;3dc?Lͣ ƒ]ovx_Í~}>ՖP=a*uOK?{^5_Z~beQ%VՃABVtb DQH5ηr{HjG5ǚvZ׏B;|ԪB3lfbak% 2 h:ZqW sس'2Өy0D}}1{03-ϱ_vu`P{y\j?ond944&Z Vv!hz^=(F\ jx _hw]lA+s 1'E<Ƙc&Zh8"bT\[9nb)m3ڥ$-G ޭp.oi#WּE߶Ty`(#ٖ/tziY}|=ekCm([1trH둆-#x(i0}VѾnmnQOTz)e\7Z"Eۅl(mP*~zȗƽݸxqK = -}g;%=[(b=|{(/!M6Đ. /E)DiI?\TQ$A:)7RDUd=:.'c(*=Zd/#әs*fx=>/kV5PnqBM..lRVWۘ62՞Eă0{<}'Q~z,߇ cN3l.#Dʊ :HЅqd%戥 aC@b>g DyԌ <!۔8E":Y s$gK!rbG `J`2JH`#dBSD*=,b<v6k.ⓇJc Yx@bFG8 *B g9 RZV Izf\}˜14Q1k9+$ƕ%(@eA tkNɲ@6]_R=w`qPrUry[nO2qٝ*C{s}o JsʎZ¼vWk)a^{kK^k%̿y :/7?K|]w*%cW4w?ycHɇ8A^&QG6"epd<:%GNKP(4(4@t>)&YT E)rpDYl!u"I `Up' Mv*܀_*:ׁKL8F18G%x6Ig#,Ԝ8 1%x\#sK ' չ(T# JF(iTJpnJ.I7ܾN#ӄJC 5cagRRSx gqn=yASQ;<Ϟ$Tj9s.>KsBLĂ.vʖL5zƔyKwJQKuF#t%DAphIRYr!|aZD@?EmW=4|/ͲͣΪ7fއO[qkՠM=,Hs?>7N:c7+s("wԊ Sa( !C!yѴ u0YgvD:E#I =9_;/ypyo&כI~;֐jT,AG}cG˓7ӏfR֨/7<&H+NE@SK )IlYʞ8z$ 9ux.,K;<5RJ3nǟng}$c˳/6{g\ppsh1%$r%D%Ab`rn>!pS |6@onwMwPJȼgJýT0h&|yqr7d@ffY7jxgPd11MԣZ!N5:*:|wpWwCqˇ^6V6r4mL ;(2n"NIKɳ>n̢8ٸ]X==Qc`r<|f/])%vU1qTɊVG@˧3tx-8lD5˞&^;úJkƺvby_o[ox-򑼨Y[q%Nȍ~nr#ow9acUO3r>= ]v7Ab;yIsȄzkw_TGҙk/hchI+!GySs|\?+U{Z%Kcg3jZ)P{i4?C&`VQӳ)O}}tgߡ_diC/I6?~ozAR ~lRSIMD$5ZLj=Uǣd1%pݧ4@%/,ML \W }#Uk}QZK&Nf\2)3Ka/D iA;GKj≾b}6NVvmz۸)KvGXЮޫhHOu펡.mHDV҉iDVS]<iݔw_U&tf*sPoT33ty+[ׯ*/w;i\'5|!sP $])lR!Tj,U)9,]~KN{bPv0^ҿzO޷K@AZbBYJ٫(c6UMd 8m:y~%0ߺh|OM6ydҼNJAz/_ Ԡ8nJ&{92HEKʑ"V"r|NT񘓕V{)(k84tPqħ55׳&9Oܬ@7;2O~/ïw6Mau}RܦҪ b&XPEA(q\DDb <ЬBX>iMec* Qgd0,UYQlk|T ȞmF@؋KX˜#)Ej+NhT@ )D2f+QgYeP{& C`'iِ֐֐6&Үa#˂f>E"ycjE:|o[?>)EsvWsvzsoJvr}7J;_͜/F+e=חOݑ0:BKu9jT;ۯ7wW9vnhoXIҗcjkB8ؽKAIxƗ&abH"z $sQ9FE[(KHjXkiN ɘ()g"28 cA%噻sY`pg ^Ƴnӯ6VWYړ_Վ?OcY[>}+U^dVY; ΔE6@9;I2Fb(u 9JK+†(} EPhl'Ml+:7s= .Tgr:}?eፆ뱜 9㎜Vi/#8G Px6IB+.@IAlއK1wdpvJlw2G=yX>;[K]wr#hӑ w|ݶ1`Gtmek MڌFx|(+{%`l4x7, gW~⌋>v@}l^G,C Y",\qwuq9-oߵx.QWsf GtcP px kڄ-#"!ɧcŗG/ \qwXwRI/w![::}֒w.SiF:5"H&}F xb=1Et5&Xr?>H HZY nn*{' _4]TMnm6VgA3(-I*``"POPhǖKLIAN 젰 @v_gHPϠ7{IQyk!!)p81($9 A 0 C[)$ە {sSj3𱈕ckP†D3Y\cbFBZȑL*A}B$\ !!1BSAyg,\AQs0MRq* D%BWGVCU= wbF'X9:.cj%gmЉS4ܗ3V D_?DFf,je>Sӱ2on^,ʑJsj˘Mv6=eba0 )aPly yѫ&`./A1@cpq/ 6j:uwzb: d>V,l X tR[.-_úuqT6er/#eTXυOUDŶ\}W)ƽW9A˜0D1uVaȪGSH܉xm,*9^Qm sy)|,U>x~v;⪮2ܩ/on˵rCl8L|xso74#~aP0ԑM Z놣?ە(Gj1wنq(GE`$Fmz#Y:.80 Į.VNl/`6*}qtS,Ļ|6]W/_H\_ͯ/!k  +o_^/_/.Y^D`^ 3WYqa_e ?GCuj0^54ФY͸)7qKVbcNj邏4˲)EMgU_u7kz}P͓HU^$' TckoZzeo{dBjKVc)L81MFs6T%/S<_;̚IG 4`,ts+2 >If.d@ɑeIL:\45킾QM4t@ctI7֖ZE:Y=jgt;]ޚ/ȱ#7ʍQ+Qku(B/UR 9|M+TBǛ\3B9Xqڧj^WhvRY0W`LT53A7aI+@H` G QS]zZKE@h VH  9AFKi[C!ZBw'uN+ d4 sbhxl롫2!wި =BmI;:v$1E$~0gS 1Ϧ@bVS/"H $"07Yv\ < m!)*b$ˑ8' \+r2MdbCB9bCOb&PS Q 0X*gTG%Uz'B j(՚3={Wӑ`EpvuRr&] M??ߺa6 9pl늣,{vXӅwWdܛwas} *v狻^Ux07Yt6NzSD5x}$uOvqR!sEbC? Yځb,ǁ9!XKMZpI{'MQ'=J4uǂ[ "\ 5"Tq5)AKXhC)P9ZL=28o-pS_[DlÍ4g+ϭoٚ&ɺnT΀Ec#GR19&wb)sD̼Hq{hOAFʜ2s% ) Ҥ~&3'~ fka$׎[`xGQмV.`м{1u~"4}Nyl1dH "fԐi~Ǔї߈$]d;T~V衛ݖmZ^gݖ (HC-{nQXqi(((.Hl$^ $sB&a(d`'P{aNN0QyT8)x qL離Ζzl) r],#YQȩ.j1T\"z'T 9}Nu5ag9*sƞ ּHQn 6KPjS6ʦA;402LY.GKV{yiK!!9#..2~.\:B+si,%,qKqM鹟e޵b۳Z=0LF2EXR.' L 9 \5"@*/ T 2UH<^1D 1ae3l?O FݿQx"t;u^0̋b0 #Y/{H}CMH:}YCKcTHJT/")9"")O]Sv4s @E*D "1|k0}Kf/VU$%<I!lo_Qf3 $Xe+RvR0:!z%ɿ,{e0Ae-&,+qI=P6y4S$Ǽ}_"L*Vvp/^%Y#s?3T-owo~(^V|P[#wbZ"UxUML\qmcCOO(>6ꊼFC"T "xRzr< !̸f~I EsS Y*MHIHp9=1BIAdߦ;˼;ui5i*'2b$I2R0˅LJ(o SJH*x*3 T (QR)&MF MDwe5uD6h \3otaЩtjyr.Y%)V6Pb%Y)cBd&iR绅Y "\9| ]Q%խ`}%$Dy%,wL595D'5d폇^\c[xQX `4{k{]~^Fc)c緞}f85[Z]nejK:}Vȍ3ŭ@hs׹ftd!eW] [غn]LZo?쩽߼eزŔ[wiyx{ "5z^hrfO7<ι^D}Wt<͗Vf!bΚϚ֍X-)ۼdf'ڡL๵q|sM&ZR.~Zᛨ3@}a2 ϙ+%Lo`el=}B\jY6S# !hdj'&*E@ DaH4HO<=AE*(4Gc>4DE`Zs&)KS%# Mk)W@}8DU[Qͮϕ \vO4yi6bʗT*난kíh'G%-AesDcl06"SvJvΦt.׳b-V ^D]:z\o[X!سIɥgS$SO>[rGI䦩1Zl;f${˼էuLR(5 IoȣRFe:%tFXBSlR K}q "u>cwwDZȰXz~*e[tEkX3=gʕXu-3vul@C} p6[GD1D+w<.ow泍 y?"^V>V陫X}PBf8,=TL9\-@\σä98Jt.ɧs=İ7~hˈU*dd8.i-cYZ3.1*/ךWf ~h=5et1yOf/z1qP `zDp;Jp Y{-ơҵte|*B(&kFaU/Az^w4nu@I >qW:(ᕈ٨Eblb‚*H$hӄXq[=ZW$aI*7 hL@(8h$(?p>&I@d ڹ&$K+c"@ RpjҔG-XʂFE5Y&4w#MVZIk%퐒vvh0YlyMN :U=ЯS=?Ё8W┳_=t7cNӅ^|b)[CBq;PٓPݐp|;}-Q1A]8gys5(B&S#rT*g8P 8mL~}5`O7_kav0 SxĽ%r-x3v"fNi,bjՍ44 z;1G&l2c=ԬM1ټ,TZ0֙&IvE_N'.74ʫAQpI{߀\ ʌvAІ`ybFK%D!e`(Em1r٦Po$zЛ힉+޺~/Kɺ[CݞMŊ\~Vth6:n=s)mBPg$т:F %`Hա Dgo=@;` :RM3]ez9];rylU\F 9K-T18T ƙㄛrW.MDAl@N"NpoydAE`j4>jKSS36x'Sh_w n(дʏwPUu^ClXջ7*^r;4.TC T@CQQ8+'bDm( CH/:~P([Qhƃ610i-y BEI# q"+d@@B)RNQ?x,ZNI\y, aRd뱄Uc k'HʨA㍒+n;w*rO֞[͌r<f1)L-GК2Mu5p*6SP$95Ay hie 5+HlsPQH %WB$pR8B[e BL~IYVgyʻ9jd^ݺ} ƝL7J*|/q/[-њ}osu<|\~7:ܪDȅ咑G;ۋ] NΊ#F藸1E%vqK.>i9Fp{1* blQrU Y ;WKKw:T"e򵽛5Bo^on/?* cx|wj!g?R+C;q aTcgiN֝Iň[u<|zE`(sSz>&Q?z7.ݼG/n!U#IZ9GU|uEfB?(`^llGcvo TQgjq: 0H5>(X_'c>bMoB#NxA%OYmrpһXQ8BRYe@1&nL|=^YpFC`ֹbOyow^A>ɠV4(4rsRTNT0MS۩ҮG;iɩJ#/O+V&6~8Krz3A!/ Ʃ&!Ra1DdJHa$h>~nsZ}!`EПD*7{L!6 uIgo*@=ki OϿv\B^26_}_CgR]/ C'g@}3v~Au{t\j:Gի?_9uG*Cl7+q'/|MHooޑ傘9Лaʋy5$ijb c~p/){3{g7pp)ѯp3>G½Crr!G/5޵B?Adx5~T[ZQLR}TϐEqHJj<,隞_UWWR;,ϣ;&5ep,4iș]CSj2 PyTPa*v6z|?vPײa*RjEkn~6mߺ4lًeh0n{;79^BþS'=\hNh"N}OL2Kۥq0>G睙Pm hH%7PVR66*uf~ƕZ'sKhdboEyϑehuImO-mx[vKRb^bLޓ]ZL2 L0R\9:&%4DEwJʏPBÿUqז5(?&oD4s8LCxɀ&jBf"ٮgY.p^:!'&&,RD6%dɁN˾"KRiDeBࠑ"0Zȵ*1HbrΕ:{E\o^4Ǜ7wa{ SCB LD$Xe@W6U&%Y144OԒ&mzc XCNNh]WǏY .i*>y njA9΀n"bgǞ~_G g˧ן4m*?p}yQgմk?5vWv=R^3.adY7$,cjndA;-۹k~˟LjÚWrIk9%Y {4kC EZMx__1fǘc%˺:3cuLLqxZZګɪtQRaL v" i2.[`C5KtRܒ tvaM-VӄMu@|Odcx=ݓ#i4 Lд!hh~3csTʸY& n 4S[??LOT/+*rt·W|q)T)mS*O "WR&4p+J ~OLHʝbjg43Dxƒ3I eR7ҀDcBMhYh୍c!`ѨruIKE޳%t0R޾5yk.n?nsZ%◫TL`emr쌹uƜܝ1tzϸr}4E KI҉`UŌpWyV?Q/l=P6Sˋ)**"% p))7"IR,mT%20{V, m.{`*pzwv5~^&Áz_aЛ`POzI%~w%w`e{δ^^hE̟-6L0q>4} _[;GJzq0>_H睙4 J`+ )T,e4RP:`^zC.,޾h)N면ꨕNKEiDK^ ƒUJa)AҤ.39݀({H6Q;O/~ףIu!"n<G11cTQ`H&XO`H&Wc s!Jko00"T>\8ifܽY A%$ӟ XX`J*+a҈(eg-wrg-lW^Nt B敉&4}m(ZS\$Z;+Cgr]ʹBڪ MߍA)3UjX6lao 5j`hdQ#buNw4#$ E婑#qڈ,KQ\8VgEqB}2j͘ ;d++q+p8ImYmG_kŦ`k'GVdxDSMoxrPg !s>Ӳ1{hbƀw9ݭEW8j_ E1X)E4$B>=iL=_F#HcIJm`FVbQ$le*ʈ|2OE2 1\@$$06pQ9)IjJhh@-qY YI_MOp˥͇$4ڈ *HdR B=<FQ ATJHW))r:l.j1Jř%8$NhtJ% 7AS-QCQ:]`hB}\H'$rϼa:*e^"9'e]jlKJl*OR q.2)sΚ0AD+OHK []NF0Lj OJ(!Dd4Q7F-iJ+I 0'2 ; hӨpC$_D 4ث*Uh2fޡXyQ[/M*,mJKxP"/.o`R}^ mߜJ8*bׁ֥VV\eemvt=]k7:3е<uũڶ󶇷/&m8"$)Gб\e UpRH@ *6˦ݴ\H H4LU6&0x#,>@ 1P [(OaҞه#ͯaf2sqknV̔J`<{k{}xQߏ'n%}f8ft=wz0}֪tzog!76PY@hs׹fCo}%Rl]asi{Tgy͛{]Yuز–t&>yx{ӕz^jy<}̔Z<.c~OGt<6ب}2z)l>4g?k+lu!m^rDa,f_bl]K[͏3:#\Xeq<\ЬypYJ  ..T;At(WӠ:s횛'D7ۼ0CWw:iQ83CQWTj#O?'Z3rY`+MM,-7ݴ\0}F* ٸ,PZIĩ,%--+EeO6W?|4D 7֋,%8ꕚj7dt\5:]|CF)NzhvƮ<k|6~|/VԺ8ToA,o}эR Rd@ue>v޻ (WS}'sN tjKi@]M^ӓ;>v n*%!eT  &-2K_Fpr0|msKҡ"$ӈLTDH/R{9Ea|xE|Ѝ9,P,mrFjL0OHe*Qϸ(%N8e%#OM^)nT}arjɗluYIoo^mMn{vK3Z=꺮ܼ,A{޼`rVe:T\ u6,eO)Rq{C8SqI٠[ B'e n)tR B'%|zuS_M:,5/1/_M,*W uUʵI2)ɊECOԒ^%5 p#R Pk(q,Q&Fj7 \o>UhC>2* 4P#7 <\@"RŤ5&&ÀJU^)Ե0z 5*D3a0.0 xč$q#݁{%پм ')M^ (K(8Ƃ>iė0 ݰ؀yoOp˥)ēF1VȤF )`FȤ{*y2VXS)|kqLpolg<T&F0裶4EKhVZgSR,t,EYyLRZQ*΄P4sL8$NhtJ% 7AS-QCQ:lK#/T3Qk$Qf 3oVaF'e]jj@%N1GpL !xjʩS Dtb/iNn[{y6ݷm߆dPI'1W`h|Afh|{KS~e"!mu?2V@ED *1_2Fϫ"H?Xgύn΁@ -e2DEp6h)Q)FXZ[k4OТۡE?qըn3[~w}k~kkO=U2 {@#A=*41feo٘/e3gv"k!ko]~|w"bvf{4hvY~( ^γD9x8`hhU-t~·(3-yHAג1_Mè;m?w/\Nmv=lz.~ .O`|Yb7'i7TUQhH$ҘHRL)qch%jO8KNPXv'b ?sb~wca875NiaUMKFN`C..hN=JO'8\6oV5:~Ōq~:cy{˙[|ˁt6VϙӆiJqx(hdv ;6q7Y)fLcx\ͬ;B: :)k=lT( EˊጷI 7`]h$0CN3,@,chI4$($ ,Eנ "Fn9#1!$ar;< s/.l:y_u\=̹)dSujoY6?MkC|{\7(7wY)~x'76*xoe0K`i\Lv ?'TvOJg`Z2E8J3:I8]W )/f_W{BBM׷mG_jaF{8վauT靟=M/άJ#Q8+~4'$t{l$Q:8;\hp~yԙNz)0`\yշS Qx9si ,X*F9_{?6Wl( HiA$P<1Qd.j%$Z& DDzRvDό6'rrd)ڧ]{ Y1wΏN_[4ҡߘ}]>dAR>GsmA $%8f8XA+L1VwlUOSY`Ynr7 >n=o]>ub-#cVr.'>0e!9iUDo8F&ѻ{6c`lAK}u d7@l'C>%)RP￧z(C=,[aUur*j_୷<z;}^5#xMK7<˜9ĹH `0NX) ͍#+?td<#_w:@8UhWன|>/rcm$hkߣ6;9Nd m.SdVGEpWK0Q0bb;#M8AQJclc_Y#UVɬw}C*(BUL0<͂{6`|݌ǘL(A8'q,6ZPu=)L3p73AU3r .(CR08U֖aw]cyhwqzsW`xY$ =Ũb`AP%Em`18( qPIb:r&um;,0Ů&1 |dJ <iWq B`Q2)@quۖ+jLD98"VM?XJ,kN*b<20GR\3ǔm$ECj:?w]Mu5mBHT3㢸qP#1j"e>(NZT^:$v.!=%VK `ę8Z!xJ[)}pi`-*y8GR!6%Y6;?dFE8.Ss~6]699}}&Ϧj 7py#5@cJel+}Ơ8"` 1 wōHFul?NWqGSC=lg~`FșPŏpO\%bTm7U{]ԯgZY?zv|5`d1B6[̵~-33qգ% 8c4 3e 񲙮Ө9<_fY^ц8V o7ty?g7^9EtuɦYAnYsS @G>}[b6)S??TTN Dž\:~?*|y^'T_nִ/!fKjg+~b~>{蓮 0{'YmZo̲ӯud,cv?7˒(Mn>mM Mfo25oso2js uq^-=+סb!ˎrG)@M6W9:`Emz~P.mW$E--PmRF9l[lJԇq[f}b4RG֦!vJRֆ)wAq#O N(@ax™DT0z ]ʶ5u2fO1IXΩӊSRn[É' LŶkgz5i W{ȡS\g˅M%rUSyJ͸`ro9c |$uB@SVD2Vk9({6\p-bK4f=ϝ˥߫ȌpFIb\f%A(Ϩ8xE53ouVHMEa 6K\ՊSwu`s `@JN֖A$<:P./>sOy{M% ':=_ sM/N::tAk6}Y"m+?Kx0~j7ك܁,˯l{\Ѳ<21܎{:>72#.ӸXӞ+F9azU/t_l}e}|x;6 7pZ+E AȎ?*N<#Z9P7 ؂,*{CDrY#!Dc[V[세=)ZKfǶe nm9)xh8*!g'K_kA `+֦!LnY3̰kWfv]Ү]i׮kWڵ+ڕvJv]onwo`-W:Z-0VIq9~9|Bȹ1mc`Tp'υAჼl.#RZX" q rsTԝ]HmD͜muKBkJ=>XX8DL!M*\bAsX ] )w⹴Yk.;|'=~a}??ϲXа=xފ;ީ7C9^*k U nN3=fluM%"ҚJo*|HVC%r}{d,(oq)){$7'zڹ:-wɦY:lDPo+ \#S^Zk7{bȇWsLc:@h~Y?FBT2Zn`EY[N03'Gա|T#g@Usl^MxY0+^,,3^e=?Ǩ&PjXß훗f(!8&B U/`WxtWU?,go̔! f^&$f/P\>SH]Qpͼ|KruQ1cnA"K\4-^ZBLd11]2΃vǿڛ ,\hߝ7&VwD9v !253}SzϘ&k ,aqIz 5rm^L/?zWδ_7A |~[mmn>;A`AQJ9U|wDEyH雷}x|^l7Ggo#6,(blJ9aW&?u`zvsH[8 V~E[|3lo.*j%7دj_Y ùT6F)X;MLv[rUHh!Ka'գ@ xM)+K 8]Z0~]KbR$i6A,&Mfx('>%h5:r F:"(Q$h'>rtriM9N־|m-g=QU7_|Q5ŭɓ? ڽ}.i {Md"Ӏ*E,9yf>SnTZʕgR )L $Nlc$9^MG7GI7{ #E4xi.E?TV\ jd҂7V3)ٰ"x&|g;"VyAI4@֣`N،x'#樔F]FoVL4'VCl;Jsu^DI'}=UoT:_}lB3UvoKl]wY("Մje^ia>bC)J` 8pA " <nYo`@@ۛcIn 7NFbܳ1 Bz݅(^=Wyϕ?\C詴?̫U*ܵ?Saz\F8s1ΡD#EဌyPnjc?W@%G$s9ni&s,zeY{ȑ+B-0mU,|,063/|ھȒcɓq_[/˒Hm[r:AfbX/VatoY;3]lvk6JJ+UnYsUcJafkG@qW",,6`#^)VIy=yu~SEH0p\w6_/[K2_OBM&:;сzzhuB8$yyv7E):ft`V홶[-WZ"Z_{-eWrhԝpئ6j8dhcr y0,+绹vC] 1-ذxu׺Co;&w z X,ywazƒ%XЊ~"\w[;FW{”o3+S sd(tszǪs$*UҍxY.9G.#8S3rf pZ& UuuʸcrX& 05{kMVbR@R\ +yu %'37ZƨdDWX}D"w)wgQR[Iol(%)'\h-X2IŒ F]E `=꿟ם eh%70s8ne.OeIW",˽-.)%iːd1%LQF!iǝOY@\f&F#Z_qdibY=H$hT[%ijԅ,e$A@µ@0# $NecH8'9oPviGp_k"=O*KmV`/Y!fp 7mNGCʄtXBQĬ3; Bw4=ѴFFJFHULJg9e'|&A-asJĒlHn]=x @j.TLh!zﭶ &|$ ]%2:Slgh]XVG.~N_0~ >iSBtբE{ɤBb®2ġazwVcMJ1+A0k~񞿽 $z[R$KXH 7=&8㒉kvrZJ.ÉҕߝNRVwSDU9X:cӒZ\I7.ÇT68P(( \ס)asx1]D}?i_S/t|u}RrH% Ynx 98?dVW&}sT0jo{s5K%Xռ~b`\΀r]rs.zofu#_;#]F|yEbyG6OƻRN7Eh>x䐛 y$׍nF5uT"tÛP0 8'7lB3NaSh⏦A./FOo>ᇟ}|xo?:qF7uPka I_? ⏗PnM=ggssze/N?o7ӽ1vok{q ߶ڰCKZ9 ·ۜq<pոx_1Y/LBx_g]}_]Y֦7GCɟPIZ(*IhGNba9 fIY>*']Ȓ=hMRvԺa]z|9Dځ_0^HWҧȓsA0>ۚMy'[Qi9uZqhbL6/mŤY5󑗾9(ȹ4rs\]|KL_6TB@DI\YgLx6+( ec~(-lRfث]:HeU3ۭ̦3Us@$-xZсi'r 0Haa3`xjUNJ;T 2w%DESuqqQpyfX"ࢄ'i],ŠmV\gR sk!{?52M8D\D"M$Z;j1+w;hlU}_lJ@H'p Gc .J, K)GnKW%#shT;8XNsm.ILq arrQZ:T90n wJ9@1/5{w7hktZܠWt i`sFu<Ք @kġZ Jdk}RK+nh R&g fͳ$͂ b=g Bz7qӔ4HQ|M30 Xk@X)_>oyZb7w9jfІdCXMpAT Nk4 ;KYn>88S쑹yw0e1L|7#f  I $3BBtu> fSN[,E-s'(~Z3쿥]򪅴0  70|,gBIr{`Q]%f[}d(eIRd 5[@Ԥ:!uABMAz4b1(dfTv*) -QzDNj;efTs+`/N}Hv~DGOBkNw*T晈P>MQ,Ue+[Ǐ|߲6wu/HU*W*0[yJxLf+4IIkTL`T,9_ T*]yuMRK&D]J>=P<o)Օ銈B`9_J`xT`3ψi}.)Q0mt$ʒ?EJ^Y"g9dJ@( "{MQTq4wЮ_M_CV-u4H{ĤdEH{$`( K.[k40/iPvmK]\߄S#h)Ҽɤ~DuVȤ =Zbg1{!1hd3مp oZR2;HeBr}:=TFYdrXxT}t90IH`h*Qw{Mj7fef,jcP1|!!haՓeQpܶO~KTl_ThXc=Z1m=Yn-zx|H_Vc}4PEԮTW9MH}|ί[fm| ?֔*)*W0yhkGDo7Jl.jA8kꑜ Nz!J "*9[on R /ޏ<ܿ!:Ծ4x)RR:qIG #N1|FcOyۧק?՗? cxr .T:qK%y'f-z^hJ>MnyvXsf#O>aYo tӦvZy9dQ\.slC}sAst5(&iK#G%p %h-c-PZ윶ߠV z7C! Q<_L0ugfĒn@cePBT=(ɵM`iAZ!̣w`]f!rX[\7襖9Ol/RP w73$4s;3pZohcTl1g!77kf2?{tZȠ=D>tu!]wc:U2Kp>I:p)F"$99 RsNRaN^=`Ǎ=[#aD<Չȓ.PQX!AAcD`O[]h.8E.vidLNW>ɷQ-dsS)⎕U3AgHb'fj`{Ϭ}O.J2CL2CHf ABN_ De"*y눣R"%D)s+~ d~U?,F6|!s,E$%[>~Ç$CRdS$JpWoGAynDq7-*FzO45X_wU~! ~mE ;SOֳ_;b_h8;Ȕ !q47KD@yH \ :QJ[-¼SC1.۾Hije5^VQ|bD).+0o@]JFEMݦ*:&t[M=_X'TN!F)=,()Q(FL!"Op8!T7F+B T&{@3mO]=8g4WOKt v+"ޓxwvx[h%q:9J@W7y}]Cd8՛kf9aflaa7j(ݥO)ҾKjDq8Io?0N*d[zs$ NH )`Qkp&MD/7"Ze O%!ԧ$#@QTDFtT"S$#$o-GoT:N)}aT1rֳhz.jde'6̢61A.A?r/O$ݣBRy5Hɬ:Hdq'f㩃DGkS0ъ4ȘvdZs$Q ^}B2NIX[JA؄Bd,F}Vb+c[Yehe^eᓬm_7lLNnH<qo< * ~u_+K]3HJlH4dFs,6TņDLP &1a02\Xb#gv~x\.:EjC+[[BT%z T^!P Hdi/ӡօ Q-lTό'mjy屧 rD$Et۠k/DLjM L&ʌdVh5T"=ѹ)N0ϖOVg{3]:6K!tVayiKOE_cdG囖ooVv|7As\;0A|ZxbZ^E˗5n9s g2sAn9X-S2S<ZxnL KzݫĶuDhK- n=GYh-!2OO/z~ <1Jcr &Vm׵lOMUBÊ*3 㱋{M+Wܷ'-anb{&iz7i:n44^6csThL}سv9k,TzYg"J}mWT<_6Ƿ7ǒYlmzᒲeacM!?nCjϵQxlgI]C탉T"׌T"Gk*h4>5/]ś[ҷ@tHZp꺊KΔ<'<!{VC2V\9X T2<ɅDdF˵5!H "Q]D:A#Uƈҝ.NN#sG"pwAA,X#g:?^Wsl}dϔt}[+hp;>;>yW+7N-n[nR!HAU <QZj]]":Tʸ1Q$BZݎr x0'aѤ|1ZbW pa.:ǧΕ|_bgjW93?q^X8tl>jTa0J('솜}Vsݟ(bTsZ7yBWj'JJe؜pM2g)bXj+x߹_٧j)tqq>9|P*P9ot9G 1M 㯷O\t K3ML.&FX99>jzc`ωw|5[n >On/ BF8#]7 Q0'dZ> Y̻߻MauwՍ2zC6q:ʎ8<0`oG3H,6 &~~SXuSk&7';{sO(.O\}Oop8!$ZFeadog U'A;a~W p'=[άyPĘ'柛OځgCIq\M%z:ᛌJ>rø[Wz%:"D_#f4#v@%᡾McMԏѣU&幭.f3ނ)1aDɁ4RGƘJRqkLT5q# NpF=@I*k=.eۚ:'łT$"M)9xbkbx*v!tx:kgt5;/>`ܣ9~i jK2T]_|M\YBX+!TٳY9)@hOmЅm*1n{@, ɶ+jO^:#*r()pCLLV2T@y !+p{h6Ѥ΢)8HІHr72ܪ}ɼ3& 'qл z)u3Lj[2^5_;QO&y>- 9VPˆ{xΏC׆3?+f^lg3G{t>}g}; ${vT8K{PHi[ a45 O>?:_H0e_upPGJHo*\4~%-S+~䍮17 (PGP! Yہ`,)%^MA('`W2*[aUx߂dDmF'ࣅJ)O ~2΀L N4u|#)_+i `Qx+LЖ#cyc*BU"y]RȬfEhV[n#ՖwITG*(D)TA)*œw\$-j{kz`r^ʹg>޷kf[zCeoWӦc/_7_R_AFP!)1p 3ItT5D w@Ң({C!9P*F.֠8.it*x-2Ub8 p0صՠSzEk;!p& -Y֢TOܮV_*$2u|Ew@U9=ʁRi '9lE-!tHءwNt)31SPe圓U!q c=95/GoEP$E7xb)qNSŘXi[֬I VP*g<U+PbԹtD , fRGa7^gmv E% O^Gz*rGmFKWlW޷9i9x8Eܼ(B3!$͉8I(µW)bTD;\P-j GbAit>WUtirO^܌)Y &TYg@%פrkj;Q'YBA{N{#(h̨(%OEƁ(VR 0N){ۼ;7,_mX.75PF&ؐ@rQeس42mƛPz l ];B׏%x'#@ )oz%𚠀z窄*h*5 ةW4͋- .W8r&߯@z/mO?jYIOٻFWٶ~1I6/s^ԅek-KlVbYduR5dz;I?cx冀x-zSgjWo"7zٻ}t7?Zd75T~n(+5n+咡V|ًQie}SwVZ}.W %$Dy%,w*,9{fgv4N^`hP5@-!x\/Y^cޒYROķ5sfwwz ?ảֺk{: 7emy;S_!'쒶n?fO9wL=}y8WGV8sn[f~{򹳲+#C.Z>v`3hm}g_1kY{-%k.1t_iQ_3aC,5QsjS+[ %&H\\  m\3&VW쨂+8be\\R !I]2bppS圿JA 1*WppWppWppWppWppWppWppwP"жt)lJ7ͦt)lJͦt)Rٔn6FfSdt)lJ7ÁjexO9H/+1OgP W>N3¶ϣ^ft[#zEjتXa#xFyOnXrV묝?& Fb0יc"3+a3Df*5)_iHgl}>](c,:Tn%D~ + (w@%ܕ6CN`zM`٧e9O <~^/,j~mk |}&ou";%!>eߗHK DuI O9^i0L|8i50%cxBM ADn8m-OKG"H.hHI 2b "24/cF3.1*/ךXgQalu'b)e(|&/lޙ1\bKлCa~@4kݧ3R)7Q7@LTȹa&E NDicN52:Ř'F:2)#uLֳ UQ*K֌fr]g|Tʎu(Mr174g>CRn_ـ ᠞h*FC 'Mx"ZBD*Kbr 8 s=,Mt@P+0uBAMLXV k٭xWsڝqǹhmYv`PUI*7 t^hCBQ D A1`&(P}$#Aǀ^כQlT"coeJ!SFwT$HPW\[x*:'jIb9s{7%+taX 77Y!65.T/a dLocoFK-& &*|$&1 2 H%-L%N0p5@sIk  d0]۞!p7Iӟ3m:l8RKM\"7EnRnАH1wAORL(qc(%jO8KNPXr7" 鷗oF;,QMr|<[#Z8bwMvr rE27Hf"5ϺHh\O,XE(,[K2ۍݽvVwm]E hBeq 0o P9ixtk>pf8w"V,L\x{` oGw@O`i.b4-0<- W#=@CpLOx3fTJ=cy=˞VSw5V6LDI>j}WnA[A%SYQu9q8v8 6 y1 Pf 6<5ZM, $ (%@Q5(jw[`L y>>KggqK5^WsϷc<z]'+^BW^sr%p@2D'RȣRFe:%tFXBAm'|o(כߖ(o{^ЋfSsQo㲅j=Bs2w6.rM@^>G-G}۔q`%Ҟݾ'!';#ןYBf^咥W*g!/۪+_LGfL U u,[7Y?Oij _]|q[c>Ŭv_Q<yS"/f&sBs103$W z)03VxfT(0_fU+ͮa s!0qfdԉT0Rde(415څ:\P] {AwR?O]*Ň- l=ʼGZ<.@K-T1~*'ƪǍ3Mhbp]8&zDaRkt;\h\8AxtM&T&F/GmiDѱ9{|~6nR~2髟) q2MrvG~EvZϗvOhZ|w9>^JJ4D% u93J.9 Q9AҮ$x^0|3pM.\QV< M$N$c hPSހ2Kj7`,{<6 {#YJX%9Iu<`QRx ܩ=uZ{n9hfH5IYgd*_FlMښ$f8(*)@-w* D-'1P9CZ@Yn#Xb($+!8)EmR!zﭲHYBL rv,?<]jd^\'L{; Q|q'_~TP. q sosoݺQ7yy=\20THP}Xtr0J('\h95U=ExC[ͯ-5 _@JnԢ@*mْ6K9FL>r6vb =7 ;x=F4#O7*,گ^x4J&V |{}jFgcBylx>P̮Wsf%1( 7umWDxߏӦdaWAH۝$mtvzYeF+W1e=閇ӳ(ƻ2lk۳u>KdOB8gzȑ_g6Mrfvf/;x):(#xΟ?ŒdZI-wS"UX m=NN:g{_P}xO_|Ͽ_?Pf>w L!yM5I~Xzdo]rE1\ދ3TUy'|4 Nw81çe Z?![<qȣEQ'Cnxl>H}/JX~1iyWBlu6:x\&JzIXIrB&'KX.$N'/U[ S[ҥ6Z^5zeA  TpM:ZA0>eۚ|AfQifQ;)rӪē”IyzrrƠ\չ[Rɵ*jY4,mDpxX"b/RP}%L'S!tJi#yw==ٺ=բ=Պ=`o9"R,cBORM pE"fFjL0Q#E5#PV%jmMAN8e%#Ob1rHb9م/땒A }ƪ?2  "Ev;T[.b?hA +g8Gi\촊!FTiڟ+?gMyp4;ʔ9׻ae4^N,1ٯrt* 5~ .f^>lox%xׇ|{)}FU\S,(9PLJ햇ejE3g*B;P3M3tTәj:SMgL53tvj$DFw^6\/,(H,δD ޫ*1]byE y\2H`6G' XQ e 13d巬i˵)I0ӀJ"S"g?\xn^p䧺e}3ƞK$456GB XN{*yFQ A^KtNQ6gIuc3!UL$NhtJ% 7AS-QCQ:<3 B7Impg3Ʌ7L[+Ix9KNr.Y%v./IKNsJ- ǽȤ)DDr*d$ڠ,zB%xE<6YgK߷k߱-hk`ԙOdJ{ܱﮆ80\\VJ%"j:FA5W wJp+1ILpгJt7WّCYztRY@fnzt͂%@sJ>K,h`)xk"&DHFE+=7&bZ ]\o ^\i7oaŅ5nz9># {Slǯ(#qqk9^si8:l̴oKWԯ?]My7-YxA{\P+l^Fg'l: 2nȁ9l80gr=fRvL%je|̆rK 4@$ Ѳ +m.K.k] ;P.W ,?N)½Je:b2LG\l;D”GZNe8EHR_%c%(ऐ"$tpHbbuüEübJOV^D4^FI|uZB7EŲއ(!j ,!V'S.ygCga1rod7]}t)1Waur}#Ffdֶ~Ɠ@>י4:}f8ft=߅zt=YkI$,յ@hw׹Ofrd2k 竰uլq:=.g[Vزû5zoYWZ^hG=|Ț aXWt4DA`Zs&)KSg +#gz2q)KDARn\nuP\q߀GB}Ĩsc"Â2j~[y{5Fև2޼_x}pave Nρ+~ \z,sa?akI9\p퉳93*dӷO-ja^Kk6*$വ< / !%R=PAe*O2Ed2?K?Ep,#h%7FPeZ #g?U>-w#Fܦ:Y)mBvr Z~?> }˥~W\R)7Q!7@L9X0'4H1cQɨN1&I6ϥ\lJd=II8 QQRBd,Fvb+㩲PGkaӛln3<0er7 yg *7/_%GkU'Mx"ZB,3chʩ=5WD Gp$p"#{Y\bCT`6ꄂ JA Kb~4+e[v-;!*I]FR+68b  3A!CIFǀޔOQ=06I㱎6JJ!SjKS Cb\)'O+LI?TOF7 ͏x)w|H0}6047.<$=j'4H@"P:Y%>$j>{'n@$B|Z ULrF n:/sF ֒ aZ4'B$"7گcp^fXOŬYYIu<`QRxQmNEsA6c\FYL:#SO'wmM/5HpUS TZ U*AZjOc>!P9!6P1RYFzR2NPQH %WB$pR8B>gDZ4GDI2v3'pC'mϲʻ)[5Z\}-M !aK[zѤ|1bWYa7 zs{Vn5\5radDfhqfDP$kElHa0J('\\i9M&.&S/``o \U)X!Z?6\-[pCm^}+L>2rb }twj!R_RK`Rz wVEFĊ71/7O|3aTژ{Ĉ֦Txs38vm!F*ʗ|vkinF9T L<1o۳amvxHY$wM!Q$P$VY @!H7̃J{KAp4d7oXH[G\?E0uWY,?4M`YŢGB_'&gE֎*QlmFu\҂GHP(8xo4;y_;M*:g{yN/$>~,?ӿOO\O㧟4_h&)@϶y!ޑ02NNg&o^0r8uq 5 ɗ{qFJhG؋ h/fquaw_5h І[ - 24Ԝ|wWלr˸_נ%"NEiP8-&c9v՛=&:$-=|NR2ɂ (6Hp]"t!KF2 ImXHf>ߎ1vyU+\)x }<٬u.s$WdpETZ̔uU>U=|*Tw= $G; y٩9t݉rkʤK_:LEz, ?64'BB@n3`ؔ+@0ڡ+)[ϩܫ sʺy7D@IZV+  N rP),c1hax&3 Ou1w e6HZCQ e Q暬獯[ k.$ -S7U|ͬx{eC]1*::uęV Oة>j=v܇BҞp/ 9g9)nܸqra0WR4:V5ـc΋eI&3`;%yp3G-sIkEDS\[@EQ[ ("]#rQnSi],#b9 \e `{n1dzJ3k FvnU Lt)mXڃe m<yj%~;gR2 "߱IHӀ )'鋤XL=]# t"sPԷ&f=GGvX#R gugVIB7$1/(R 1qS mo}'S&h&BӶ <6>:0&<"Ju2\S9 (*~%`0RU/\8eנ?X<:^P r"`m t9"F'@;o[6+TNr>(0E̋YG(aprZ%dz{f=09O}#rl݈c0þݚUPVaZAuw^L.ԭ1o5& 1Z|kt+t IKW P_AC.6f$_iZWw qu5 n]n7-׀8]r~(_{ ~=Onrf_wuimMw-+e}}e|ga/yno~ "C<ͼó+! \wlTWM41K%IvEP^#ly%!D9 ,2G8$`ȝ8 =(}\NAE AWep.*,S2!rzf@oХ(xCZlw/G~zϭf|zm2k)(PL%A'9Ru : Ls&H))(<:FTR(^Qs̤qd<:2h&MkKb̔/.B]-=kCmHB17B d#Vr 4]v.[vbI9Ns#IB%$N>" L ".ycGcWipw۱k뮪`PI# Fgn_ *P\]j%Q#7 $M0F8LVn]{ߔǟ LyǮmr޻NYt $o\zxa5&x^QW?*Xx.:3շyk0ĩWQiy&Ԯ$cb|~y*軫nI۫__a J}ً05?yM#=X|K4i,݆Z%g/jK%5|FUyYT !KD/Uou4>.w:;D|Qc'?_h2ki N`gߑ%%a(筁܅$>Ӳ#m3k6 8bnǷa?Hw:]s:벳.Z]glČeeY0I]OIɦ y ;Clsa퍆"iasMf5@gch4&_6/~y=0w̏'O`OeӬ0콒R7ҿME4{W Wdկo9AǛ!7mA1A~T⅁!Kb2<Ar:Z ԀR(te! KQ9Bmpjl$x5z#ÛÛDG7iN8\X4Q׾M>_UZ *G2)V%Ct\fǑ-=+詗ƩEt[y7 >4l_n%ͺ9}]cv!@p ]+ͺO'PkeK]I1 #8.K˼<SdLc\Rњ$P+e28˓6BL9 e4`W(5ZeɎy@bƒ}¬_Q!bY7KB|ٖ]4|3(k83-`Z0vƍ j@+c:4)ж }ɕMUN6? ˢ|.$X= r)[|7+le<oŗK[x}V@rZViV,gŒIw|pr||fC&Fl60BrZc1rV i`Vy) ;[LNx-$Jc+kL||Z,>sE*4]i@19ODغtGbfȱPa<UeK%Q)4Q6p@Pch#<$dTFNjYE R0z]&gp:Eŕt#LBdF՝7Z)vse!V½Un*=Ugu0֠qvO޸6ZhQʐC&{š 2SLc*e1xa >RIխ/8Z(^HQĦT:P$:`f6U9̈9% sy,R8,M;NioCX9ouV-#,sIΫNt,CUyQiA!eHVOA&M&&HE#(122LҮ6}5réS-X$b\+KDIN"VpƔr[6+e r>` Ʉ@N-J]}aI;N899lLQW$N AvS>QHvD2)!I愐$ V'G(T ]qzWPZ&`nKww*eln&P!XpRj ͍ >s˲]i+|Aea{^ f0)o]y ~٣˗KW 43UE>WX5M3,1L. T*@$gd&2q.3[-u}>\cE׬.;oڦFXS_,fIu|Qުn=odA;-n`Iod;,] Zbs7Z-f>5rjw3;pYwWgf)^)CO<S˖U{P;En#I* e@d%m[C7z=hƣ%M:)wLRn%Iu|;OXiIjV[,+ݔ kZ+\۰0cڅ{&a>7P4W'KN͆-}>l.6UfbuIg /R[hjޒx1ǒ1m3.qnhNxީk*-)BD 뤳k}-26+{.fqw?8}|y&a0i62+CbVn42, , .}$bu+d_C&o>M/mqUMG/{EHn")BSAA j !7"55Iki@ؔC墔R,u {JSO_S{(ZۂMFk̒e#+[tRgzH_!2;avOw`܃y#R"ejU$EQ*"i+YEd.h{)KV)cIlU,))\H^'mmT%20^_&<Ưٌ_QCW4a_6'Un{iK c^_*J4hQ#[7Mbb*a廘p[̙ +GWa& VR ]Le$ FĦt?;#&xJqNt BᕉK|m("/o)x.QE2LΓ𙓯$4Il"OVBrjyt?O?\QR dQ#Ξ|%HQyj$DtH6Ke;Ru΀gQPL`&Z3&*Rh1r6K8h4M-ݵz\Wq)onWt,K&2'T LYL.R,S+鱗5T*ޕ5˚)wB0>kpnhAFH # X]Hn9ka)LBS§ .}dOx=={XVc>r%*FNT ƙ:Ǩ'1D7MZDaRk4;\h\8Ax4M&T&F+GmiDQ/FΖЯr-LI?T֧7m>ۧU[4x9ܻwqm ֒ aZ4'B$"7C6JQTy РɬVI-۰?1^xx+MmNEsA6c$jԱAr; R[6 EE8@SP$95Ay hie gT URr%D'S*DsJE.}D.$gك]W'.__~i͟{|0 n4^cs\?[V |[tyK7&?Ao~ϣ B,^M/'CoŐ̮Ivb3j~th}ήمVSdW~p{6:fO W`q7 WjQt *mقxA0f)UkE=+>T8kW+zDݯB_Oϗ? cx>k? [)hIM{c|e8x1?=ltODUn.G9!riпl֖ۻ]ro7Y/"-w7w3 LWmӰiN,Qh# G~̫wpp;_ݜtijZgedlYuda.#yb8[L7J6퍚CrM?^__";߾>Ï?Rf>8BM\P[Ѕq͟? $Dy_}.*f9^q'0E؝/#hoOmHq܈m%z̫K>r˼[H}JUUaAdGA}ޕm>ٱZ>xZ0I`CsjoޥBoq'*oކvjz#a ф*H&s ow^A>ɠV4(99xY+kaJ֮X.G;G8Qn \yw-nUw33vYTSA?u.l.u/޼RXu P l* Tr_śW>FVĈ>,9"V!IYk1Pn`\1!;, Z6!ïb$ƺgaw^o&>pOC:x%!Nieiܦmӧ{%d$9ɨ6V!hƌZ CcD$yTFl wYBL.sH]TfPIy1M v+Av&ʁWB #A4K]|NB!_jidQ* >RmCČ,J#kt'\BdYpg$5mov KMs?v&p|T#~KD.ƝC+TRSJ;B Aʣ+:B)/-yRMR},cǩw3TJN+ mrFHg<`TetubNYDSa(X-^>&i(g%jӘ㾐#M׵|ifNbt%P]U9a4{q{)~Z1ٗf+C *ϙ%LO`e@lvqذ m(%FC<&OLTHq Ɂă8h<+[x:qn7{Jؕ2JrUQG!"tJh%*ɭ5jW>߽wcC%,1⡗kD/0Mq\F_PGzEu~wХ7OїaK}SrJ|q=ghϹiߓ}g?!ay\6eXBf^ݻ^nKgu l0Hg˶lnwiI PJmG=) WhM~7&N&k#.ܝQzT̻ۡD̵ߨ5kg2}jN|M5P21C7D༷T,N9u;˝^g[ 2rkWihC V-{:?r<8X]*٘FoP#Z20'dDV$t$䂆HYDSrEpP( @x͸^kbB,p1Sp qѼ3]WOU%~Uvoj-eXQ Vȓi헜]pl&sn9}xT2j$R)7Q7@LTȹ a&E NDicǢQbLޓG[.&3+cn=Dp;Jp TiX͚q=J9.,Յ..ܫ.\Pmz-CgNvfe_f@їAs6^Xc}VyH!!^qmw(%114`s3W6#8 sm6ՆBl 161aAX$4N/EfGød\.Ǣe;ef䔖6+ceL3%x51WDBCQ}.x#s QA@bLP!"$#Ae@*a1r6ڨfx,n}9Z`Wyzѡ ן|a(0KNnRnV )şqodU@;-ac~L͜Zh5kV ^Z'/!5$ eIh\X:xas u2oL-,Ub[&}@ DmK,ӳAoI?31ixY#WEWtglX>^Cl;Y8l8pcs{lIE C[h6j3{oq\ wv1i4y{5ͷل-;ef(N:\ܞ$:_ܵn( t'B@.H`2f1XXƨђhbI QHFX.AQDl=SКSaV6:l_"b)T]fЁ2`ByBPHUs^F%V M%T*\L`O_8F>??JߚTP/&k*.rbC}7οĻލH}G6m^m..Χ7UJ1<ﵟ.?z4J@#8vU9|1><^s{_P?+?^> 00 Wrs?KAp46~!m=mIۺaݨydJpT'$F9=Qi7 8'~m=߿g4G??~_ˏg8qQ8k{. [X?z?0|ߟ|7Ow9CWW¬G _7gMysaY#R/Q]kU߼k.].z6|~گuoJuMfFdaGAߕyjUozM<{NR=\;f3ނ)1aDɁ4RG'hܺTq#NpF=@I*k=.ؚ:N5IDɦQqWt6'].D5=ya{Α\(\y|r *-sK[u r:8ĕ(Dպ@w6+' Q&\Vl='Yg^=O}pGUQRʭd0BWLV'4v"u=MN6DB-MT+$с/|2T!JNqRgPe SK'Lo|`ԷY岭t'÷YBN+z4jbmcAGAG ^oyL*;qAD,C⌷+ӚWW9XEkqHH!$ .(JHc`\I ,'jxGR/QGZcDXZFP@m\J : C}Hzy.6WB\tckdelH#<B6pډªӮ 'ۂ,ӑ[r2Go,-3~6K)h9zr)k]9Okl*W3Jއ%\E14=/7?_x4mn 8Њ^3"GC?~BfœFq; ,-c7YJ%~f sJ\(0I r**K p*K)wh$n +/BqS1WYZ\e)eAW\)ȣ _=ǻi4?r1q ;TTc1'T?I W1Y_ L@ 1K6\yvj xGD"U.'@q񐷘L H"&N &D @'D͓̓4t[X纻3軖 ?Lm ?G|6i}֛{9޽7Έpfɖ)zszCH\2 z2gra 8,YAg]M^*~='aj̱enn}kj%M6~C3}G(v/ T@Hx'䉉 )sQ;.!9Ђx0Q>2t+P'Ch",S5CU۽XOn;]z O(5 I!*(QD@N 1QKZagjmrM5k^_=l[0C2]q5Yl zہzRKXC ׭¡jkF eQ %-,TY> k܅3*V [mM1f׬BgJ65d{}c%W+V0O:}wLji50%cxBM ADn8m-OKG"H.hHIl7 і"rE(@t":8̸ȂC4R'rav7 s"IRz$J")!d`j1!15cZiNmFk]:ͻӾKv|On˶{w`wV8Ep}}8>7<&HTGldג&BpS li!҃,ǐ({ؽ:km> N면ꨕNKe'%m/c*) Iz[AGH4 ^'UL<65sbl)P;pCO {%)r\ǃ]8%f6w=:8J!FK ֖R ]Li$ F0~~ 4+TS$u#A L,BChCyyk5OsL{+er\^uΖS 'qi;t껴nŧGRq`Ў 2(S\礖L3B KHj@8mD,KQ\8VgEuB}2j͘ ;HqXY?Qj =Y[kmSv޲ܒeljY`'BZ4g#"t$Lf,{ W#xv\Fſ꟩{<K-c??B5TP.wU& goQn3׋8SYg!KV-ek/'c1gHfW0;!5hd?:a4'aT+)n8s߽:PBז{U)X\ vSjQt *mٔx 'vV$sF|J"}6>Uq*%R=QMɧ*-t||4:8T f)h~z V(=Xq=|YNk48QoLō>.~\uwT7^\^;)1[ӳzl'Q^6/ۅsAHSK6$-753|>,GW G0b^dmv;{:il[tզčb0DFroAȯBOEVGz_~ yCN]ON^|u?ɋO(3'/#0+F5K) S]Xo/zUgr>I6j1o\^34U:y%rԋE|{^;;bRLq`rZAMVޤin2M6{=emvݿͷkV%eߚAfQifQvʦFV5X) ;ozMnxtG1s9\J%TP4YJi|PJy:J-4q$J) ;WB #A?Ƿj,Xcg⏖1KQ?hVz11!)w ufOxp&ICs;T  $Z\'xksUc!`8o|"g1NhNb>ߓ VGq5o8A' @Kpqܣ-#mO0 lz2Q2<(Zev0JrgQaFz0ͷF![VW$R+VTz#˫ٽZei_3(~/ Q9B߫dE&1BpDȇzUq4!b^K(U}xev1QB !^;HOzϗ츨mkk" ;| 2^>&ԅs|.=&)eRQi-K+,)-q_V3-F͝vd+UwŤH&r TRbK/-sY"L9Ўu\Vv¸&Ci"!my?ә^j|㦎o}2lSs؁r_0I-h׷&Td A-./iT]\~.>.wGC~0^&^tYR6gp\{t:];Z6"{HVF_Sa83Tk)DoTe UO`9,3t`X\]QV3tg<8Q1*Ÿ ~{Y?!bxSC*a97@o^&46uDceB?Qr=8.n>-LGhw{lq/;QEstEŞfD& aZLg$8RɬQn "[gчd溓 bձHۉ,R%;)gV]|!->^ 3g`1i\H0ɸ&3B>dv9wM"kd}v+CvUwWZlԒQ+'Z2Ü˄eCs! )PB ࢛f|)˴y/%vi-N-Zsg#];6\5g1 )v)&.ڸ{HoѲ03a4Ͷ gk I!n9,nv} K1YQ3C&] ay ,D&XrX4sSlWƩs"~L hrbNe9gX};~Pja.G5+ǹ?aY>jXy!;h8F>.whAd /= Uh*mdt&X˓ә\s2ZEuUr!oEL ݢIRe-7v: B:p*e%:*'</PUl6{뫛T>[b=pвa" 2DZ"[H*o*Rv_(/z+xGB:! fs:ͱ[42x&E4hޠELB$' KթL-.2WoR\W ZWS(MzB>?e^-*c,:\( ^EN',~\A '"k'M=3R;_V Bi{PFH4 tB8~ ;Myr`"3l*|6vG?~.Vp6ƅ_P\l6Vk+ȗpwyI[J-< ;gf(|̺CfV?[d]=rue;>xyf:s+-x\/>+yx5M5R5rK΢gw" zo$r8zWk߃!|}Y] ͓wWTM7-JunfR[Y˟1\vj$pFr9Qqmdz>KT f7!+x2Z7!S9wF&ofCX/v/ȱ!{fRF}tQ(\v4Eiو: !YҴ]yw7-S$HE:+LpN )@BN-Q>f: {HiY6c[з|9cԼC~7H'. "tN$>XecÍw:B-QbjpJ&fZ>罥Be :Q"u;yJ{Z.9~d1U}[V}LF嫃僫£Z=G~u]̷t{ U[ :9O7O[nm Ϭ?욮]x7O)mLJp eeEhu׹k3>C2HYk. l]?.֫t:kL_ʯͲClY`ˢͻ]5Oz~h-As-WCQnsS|i=/|Mᠨ{.#hٰXǚO|]i̇6/YrMl״^xÏP\]R:@v][y>EfoW۝(tGd3- *&&8a x?˖]{̖-ҧβGdҖ,ïYWg+c;ja 8I+8J<^(} MN$!0Z5I6:AI kzai` 4tzbBHіYolmΙU W~Nj(~) 8҈PE5hۛBηf];_OL V$O1b$I )`FȤ{*y2VXS)|" &76Ix%LZʠ-eFΆ*HAN󘤴>T (QRpIJBn^[&tXT6smP̚єc$*qRLr"N=ik1 /1ɹdDdy+踨W%c9qa̾q9B^dR+%8`V+OHK [x u/i#X.qc^7ްof6[ ܁r*[P69v nTyMOc)E;PBKk p"UJ"B8K{EtJTģP%Fv¢r+rnux^h~w d*9]$v=mQw UJ؞dU$ M)X[w6B^i6j^NG[\v <}/!xr? C4t& G,h50%cxBM ADn8m-OKG"H.hHIAe*O"2KXGh%7FPeZZ~9! _s֑bZ,P@#0/' :Yf=dvS /i|OP>r+2TMTqt f9' ̤h# 4R`تdTF$;њ#E `zDp;Jp Tے5r6[DBٲ,,ܫ,\87~$U77<4Ix3E9s5gpWL_H;\.KCdvy8`>;6@ X cō7ݽvVgmmW~=:3nۆ˘=oŏ:)wLR tzb X^Z[i_11c=PoTZlr=t7b̺I%1< Q faxsTʸbD^~Âr.;Jsu& Ӿվꙛb΁򪴨qamq҂o 9f(3iAe-&dz(AM7s`r6 L yKkq)U^V6枯d)]B)AjgK/UPg޻Cb^/+~xWM>ƿ&+76F{'!SNw;_؍ZK4vx'*˖ EI")OBY)/3;_et*:/1x!?JAxoB릋QmW5ɳiy? {ep *eC-uT~CJ/;3MLjzWyLnGq%|/W#GK(zqػގfތw(Gzw3EXga9_zo$)8P_L!s E XKߩΜ2ՑmSyGU6S# !hdj'&*E@ Daʣ ғ֝2q+GSIp9vL8?xw7fʗT*Jkíh'G%-AexsDcl0x[omlv~z4剑|&iaymE-v.2s#PR_jzMVXXTL NB ZRk ٞ_^Y# %}{ 0$C%9p] iN2 G y,lTa=P-T1U C֔xq| Ij.Q>p4t' '76L0LZMQ[%fQw5_xeYt/;r2rIT* %P:Y%>$j>'H܀ I$9^ EߎsGd8eC.0%OAhô:iTHH P!)(eFjy=%C!~T$&O Qǃ%WS疃m HٿB#,&ec~饶M ME;CJړO6P] mUvE1}HɕIp"6NuW YBLq#mE3c?=dmzLE6Ya'A<9?5ׇ߾y]}u?~s#Go߼knI<)쭪t[^kB2N5S˿^ ӷ'XWw>w!pΘԙap ϦNƃ8{ʆ'7Oo[ |R?޿h'*˛OѺZ>.n^xZ仺3^]* SU1u%o:xX&ljgA3GBf%IO K1($핏E F%!Vw$hܻ4-OsKPF0JBBNDGхRck&ו +0/tUmGݩjy؞3K;IfMWͯK-VWgQ麑\_zKk|]hu o&e=fKC޹FkM]lѨf>AOx uUz7Y'"5F2a@$!R t;Ig" <S DUt=u,$6,\'uf=u>OxZq>Tͧ7nX?`9usqd7"?+ܽvWN=:z!tdx~ttr>gSnFt]u$HTMA6WDC@"$MC"$JAP*gk!ɄrN)$D=t>ɺ$$ Q:D dݞDA S`.FkRD@9Fm!X/B:3nHt m$AUZ[فq[3BAO=y!㴅H~Bi Xrي(k4rMlD3 -,$8Rt5egׂ %h-/<j) cDQ`'(=tsCGLbM=+kx;[Ͼ?`kW!+ 'wiSC̗$X:+y1O1Dl7>O>Fit8)D m]kQ46K$:{~n)t1`L ^3I[BIe- ^,^bbJNXU%gSras*V%ZJjLO﹗K619}sgȈ bS>)[V pLrI&C yt$u q|1Mʪŵ_={\WpQ=|!f ̸`D`3ͨ j4yDmHBR^$ I dT灰96Ӵe."^A!Th4&q CPgx-xEEVDțH$CVh!HS2f JhJahKwʗxI t)3TS.ɷ)S!"'J|3s0ĐОs#=u8WPjou;2qϺ`ݸu_0W˷s.5ž\z)z07 귞^?\a֒|z +Az =~p¨rWl0h3kpW[+vwUwW_ v]U쌻+ZKYOwWJ՗讔ۿv])wU͕rWUVmwW1ýܕqRmyUy7*ufvp)V'\O_lP5o|h25qk˹h3Iyr{CmG׻F+j0θ_7]un4[z7iTZnSv箪v] Įj4 tWarW`v]$̮jz\TwW_"0IPyݿ0?FsL<巽?45ӆ7>~\WH%\|l.TCY_ε,MQ|'aQ Lf&aT G%D栣ғm9s, ݀T?'Fj8Md4lB%ܧ !eí,xW7X#8Ε=ڤ'6&{kךW7;xl7_=Ab8*QõA-h%?8]K-QNF1@C!xdϗM5lG ͓[khc2[+2y^ -`(Q(Ȥ};ǪTB %LV\HsPEU!ua$YFĬ29ۜTB*B,uZ \֢uI/ٹdk.r&.zo t.8fXI*6(R4C:'Vb;,}?fOM$%,R*T1)ch(^@-hV ͈%[HȠ:#A.QV>uycb !!fCbK]U"Sjgu!{jԾ:>1`},穎OkWCs  T]z5ν_9&~,Y_ G tUq'iKwD]k]c"Ps"JȽs5J{~ͽD, Rg.nMҭIm[#rxn~oL_7ߖeg5Zf걄{^wqR5=՛\8\fӉZc_'rzm]Rh0 N|!wE7ZU%!,I%*F,.,wߍ.b}8FeP>`4ʹOjO2}\ ~OΏ;~o^ׯo_ϟz;:7x[`#O {5ؚвm͔z7W8s6&_i^|,i'\nn8{ʆ'7Oo[ |R?޿h'*˛OѺZ>.n^xZ仺3^]/ SU1u!Z~tM‘M$&I\;$*>1/EƔkOQ>%Z)IҦ sҴ66p63E3 ed $D*M}a](5FY`r]2+CN':<<b8O 4t'= @YZ;E!y؞f-VWuԬ[pUtَ֚βRjlGw_Im 54B(6MJɔ. y6uU Fvw+6%6-6&*%Y7ڨ$ wN1j%:ȅ$GaՒ(JQLb2B1jgٛr[tl'5:|c;lX}y-eBS\ݚOϠޟRysr !SE`aR뮈 ݸEOxE:E#F:Eh2dASpS 7pc .A Cx$9yyH$w|G~)D5%*9*r]sL2C_Zc*cי9w3^h>ڊ7cwh~UAVP1`-~,H`^P>i8ju,`PǒVDh\MrMlD3)ũ-W[H)NmRZP$x U@v@-e!c ߃1" G@5AEekbF6>Oyۀ:w+mࣇ/1R:g>LcWA'FF]g)D m]kQ46K$:{~n)t1`L ^3I[BIe- ^,^bbJNXU%3OePƻQzڠ΅ )RC$#;%aPA̹cs~1g>y?8sfħn+{6Bm-t}GLOyF]ۻt|F3pyd K)G{ӥ\ Hh* otG,O&!e'v'ҕZ)1yP(|#3s`_F*WmqmgWװekxE,_Y: [1Rd;<!Th4&q Q1֡)3QHB-f+($ytv$} -sJ@,ra@cJhJafKw @xI )2%&kA*LMW޵mq+_q):j 9a_Ow nQiı\ȇzOweUφcGX>0CnÂW F6́Ѧ|> ).nr~*\VSr6? ZRtJ/"Es9JZe2`}fmZ"NQ-DZ 6"O!3E-EVj޼Yά3G/|0tyqv*_^fc6;y3]jul 5:06#҄BY;SE:T ml!)/S|gK 8o ׽VGӬ%y((tUuS>wہPcT\`d6nIzl2#17|{~LKESKjrtL)WKcIi-r7}PorT`[2%`c/Ǯ7 E~o~Gvہ^ Zg9U N4v]u:tsuy6\,<5d*H0j4Ո`~)79lq] w쇄^|=x Z->7?*z|ྞ?#ᷡUw_ͯ/>ϴ:໺;=N7``n|R_/M|6=\3-p 胦s{S RD: ~[,ьxjLO16cjE QXpQ; [Vׂue7t=]򟳳W킴 iY8?>}4J擒<5 e"]Ēdȑ J8T/U/j{UJʽ g\J!Se>VrN!&-s@hJڙ2v;O5]38L")êj p MVtj9E|3L#_c[?Zyx~:eի/p^\>~`T|.7gvI?fLhmVkG W^z?<]??_WZx]qWwO?<>л{ߜ\=wNsXݧޯ|o;C/[|_v4w_7u:~`G޳4o{뱻9v?{~fM0e)8}a&a-9iO*{@ܠk9~KkK?~KIj9~oxH}U*^QFp/?\Ebp W@;?>]DwմN~φ) ;rS.: &sX{Qe†[m.YwV/t}sqۂ}Kߚ>e{z&2Iݿne&Zcw0{gKzd1 Vǯo {yc+7[:Y$})ԮțO،>k/HG'~qr%~<\|ZfeӬk&;3m(By!Va;Fvhk2č%YPnU{%9Љ}~XŻ9i/:Yʽ_ʪl__65~1F6H`~OT_m]?oťA_(u%RC?9-{)tR5EHy5]O}.9DS&ժsP!蔵*e=Wʹ)ZU|/;sX i[/wJK2L~#eo5dNHfl5dc}hNi\ $j誦Tj1'aZJ=–P3c1^Ú1gYLk%ё&!Qc?X1S$Z_ZfKAۘR|=†k'%, k2֡L2:熩BQc|@TL4TZ$(x;Q~nJ=8|F\T?,I}8dŁh#bM13dyΙjg`xPHx|ܜAwT͍)IRm9wTI`UUAEQZ8#ሳvIA- эvDOѝŸvh^+bT @dTSuK͆E+xnJ3"Ha)ddbb  a}G :8ʆf@(`g_*c X<[-Ki÷KCAG\P-#cȴ}撅l+ +Ab:XP. )YlXJY~$@?aBi%E)ԂQ!(dDs#6V'iXe+R5!btoDFd{Ml6xϸt);Ֆ َE$u!f@57. # d BXfc@C=RX a PeEHc#5O(yY9:+,|qu0 8VK" U "'80)Й#CK@T:<A` Ʀ= NVj TX@SDC#)u0  ,aHl N; Bu+v3RqgCVqE]%UO U`+S LBDf Ya M44nx/`NJ I@<}pܶtiﴘaWE$' cB5(5bBE} 指'|_ty_hElB$h5M? ',X]d*P?Q\[9Um,"^V`$ S`š`|v6Ǡms@ T%B0#L˶=  BL)#,=4K p+1FKJRc@` 5@@!'d9a6n(ft #Z6-l0z5}Q8dW}݆yam LghpDҌ`0欛J 1tհL)X,Z q> aVҡ:k4IFF#4I[7jPJPӮz+"և2P[ QCW-y+Y SŐ+ K> 0WZOz״zM{}zq^b$XD0u36yhfѓRcٸ1%oE{;)4YǨ]Z5qԦI9O)mn0Q~9KlO 3F;#,HtCYNRupw+T@=@u!%L Y!͆P{u`}c^'îGc/(zAV$5SiOY"GovHCy9`Z00S 1 8Fr%Hj1| PuP)clFe#0QUW ZfEHtXCPAZ䅵I :R@r*b=gmNAk[tbI+- Y&輫'vB@侮9zJ=cfWvq7_\ld qN9+=2ʔI#U#Rߛn[7 `2bD.=ٽ.-S̷ޘ#V:,ghwѴV9DMۀbאAӦM$*N85},T^jT w3Р>U!u={b Qa9p tkST+GA_[M[m |5B=RzoԶ]6yD_( V zjZA^|S+H~ijwhj蛵b}/EKR}/EKR}/EKR}/EKR}/EKR}/EKR}/EKR}/EKR}/EKR}/EKR}/EKwmIW},0%-oO*oH;/]п͟Vh%1K} mRA T+`11^XRv*sU:\pBEhpH([;exۭ.zSB2[/d_,WdRQe(Sg.5h,r D#夔0%K 4C)˱"9@< /$̇U8hպPӗRӌq7.bΣgg'G&'h8&ͨo7Kըe hr%ŵ}nb0'm5_aXor~mGKDi8YL׏Áq^;gܥŖf'w0t솦^gy߰6ZhC 2QDK"*KbK\qsU=6p$`J"#{Yq\PjO+0!IX Ǯ֝q%<]mvזnL씖&;m"rZjJ[6@Y%cDw"IBCU.x#!(O|ML$( #(1RvdLZwv6ޤSq͏C=5zDf*iQHT<Ѕ6KY P5 'zՖ=b >8iuJ{fɡ~Ѡ_DH1xR2YD6%S"IAh"RCB;YgO'֚Ռ*–4rۚ%k7q >{Y=k3YC'/,C# y_+Q i:tؼ!t讥oA^R<c5#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#:#Ao0mu͠YG@8Xu 8qosSҟKz ${?Ԭi̓mzѻskg1=^ ˗[{J8]kf~X~ j7&叅?Sw1Uq6ȇ '7<}{m YO0}Op}K2ɅsL&aMu—"9@m jl}~oPѯbR2[5_vvsvSvSx4u`ozp3.vd@`ͭwloCwj3tYWN.O.[200_0"k:eBxaBx4.LSzիƲS]zҒZ6eJm Wlz~&B3i^8zMv;hEm8FRqI8{kl +}*e`&0 /܅zĽGdݭmɎ6s7]LѸ37k#wiN]recpOx-lY/ݖ-&j'ѹc;`~I4+NZ*k>ӟ#F:o^ijw*rO֞[4,3BsjȌUnpiziCbs & TN!uԐ Q´24tQ \M9IɕYpHrRqJ轷B_tKR$sLxĤܟhKQj٤.|-ȅzM޿f-7Jeaao&[*` *|(1z?ޅD_SoUnrfdDfў_.'So-,{_W Q (SQB9ag#rZPdd ,]f~mQ/e<V@HI-؟DJ[nq-;2`ZO ]L`n]7rih?۟zΧW5YJ)>/2\gagR+8z ,iFwcB=xݽen}vwӋ> egoA-cI ?^O[0{t=iIBn~UcXc<_d 6O0e*>\-zrUhSKv=] F|P)0}2$W9=ጃ_w2}i (u.QF?,}afxw$ʻ!]%_,!=zp|b+0W Mt}}aсG! [<ϣEͷ>5_Yޕغ>ҏt*m]{-܍!x0=b#yr`u sdV{].gbM.dNg/-E|',IwޥiS_;qMBhiYG\3粷fw^rAfQif3Qur9q2 LMfe{O^؞s3*rj~G^xyd҅xB&W_Rطd okԛa2Ŋg2^RLxs%,nXCi|He+_r]En3Ȧ$b49>;-ۜy>{,n3m2j*zȃΩ۔74sAKiiA'nZ ISRǙ%$ La] a]lPۢXgʍG\׬Y=~;~48^jr8ۣiW#s/̭ 6 Si#3Q+RG`^ ƲUJ=R@ Qg@}J0NSP6ISv^o~E5ute{]a{ꀜ0g$ABWJR*C+.`if{lk?.] LTTB7"&N̙BFDOJZ GO:$㕙K|m( /o9x.ig|ZwԿ{Jjow|4;ۗMظ84B) Clbv|{itBmJn&)#4\ڃ\¥Hj%{8mDU PтugoY߀izZ^{#h#-qFYEd#7 9FpF Fҡ!FjT5K6cyRx)O.T'B!S^$($2TM) ՐP%1}adA 0L䅫 @kVUպ&/TAΥxH+ﳦry{kYY> yikLրr;}`C +UCNB:'λ8)fυLGAB1qe#bP^5p)NK[`@ )b)Um;[~]hln W v4ځw_00Y ʇ6sпf:]QCx}a%k'Z&ÜL,ssB1Jmzѻ|:|s6 }X^>u!?$b=_dOqqIwa?p ,և)8S@n7,7i׿wO`חu>Ze\iЛ}e} Ng߆An:gvWpYm8|86Ɩgg3jPk`c| ; OX{@|זj>e֥!UMHTNxHv%Mߢe5W W4AgUc]Mau9&Fm਍i[oV/DYvA ^^Z]liҢU.j~ "$dc#hr .4I!EH:p jD"1:_(zEs\&*s2x#,M) )E/(18 U,CΔH]r0O<պ7 ݋8z= mO8W{4 i5ٝ'm?ΗMWm0jC:w'vmRj]f]moG+q2RT Ip 6ɗCZ\Ӥ"RD963==OwW?]S]Izj! /6vOy駫OtU7ax7 d%n_SWEk%:xݧ]ڼOe;(Gxqn.R',VlWeeylfU"nœ0 k[o?eBj{ roPZ^Gh"^Q 3eVҜ eRoC6Ӹ18ϛȉד$vJ5H[jTq yCmSOuxrQ XL2fe\c(RKuN֩vɻ4w] ~3VOJUsOodHfNFyi-/U󬆕UDPzw :d*s(D檠 ɶ7t*ygf?/5lRQ"pe:2).u"ǸV9az(q_R%ܡսg=a4yFtw-en[X{UHha3XJ8r&r]i.~TfgZ&ikL&VX*** kKrۗ,YV9ˊβr"0
"!өj&lI.{H&BЬ
))}C(II^ GJsbRqA$-Is-=
|AjEtƸ1x)z.+;G5!xCG?qΟh48>΁s:%)cl4_fFeRK1ky&H-o69DI0hSj64p	a'uEt00c7yqZ31ڍIG[X[kOI"-^PnShrZFXJYW!2.YBF(m|Vq@5CԢx2䚔Y"5*QxTTbdl|Y<FE-ؘ|MÌhzF$C9))I3!Q7"%V59YlT6
$wE.QgB
BYs#-iFz!?niS9ˌC˘mlLJ^ʋŞ)Y(EEハ)@4#C3L}tx1Y}bkĦ|0(JUy	w>Vƽ'(BMpzj*k+-EU"YB᎕;=,BG.~ 2iy'OJRV3k g%d. l6N
X^4o2QRKA)"
)꺓 tiAdoe610װ-׋RBXл`9?Ftd)lY飃WJ|=j)GZѢ]GPWp,1#YE|A[-Qu.Q6pVd鈳DπF Y^qO=1)3e [KHV{[o1mSB>؜D*pˍ46ft\C
mb"g7!&/@=yGmK@܁%I^0!A'm2$mɁrMAB9r`"
52+H}$5RI(^rvHIXgC_9.T(pGct~oU
sv[	Ł@5m
ِ^zXXi3S'ýF1Ϣu<ަGY,`x}hNhoEs&ng__ιZZ)T݌1Q)
!,uGgR/p=t)WT{\+X3ҏrCmpUeX&m7y5t.&-$.YvUф//דṇU?%7+^	ne%Y<Rƀ(ǂIT"׽،8mwM}lhs'@<Uk9+St2f2!ʆ9DH";xy@s!%zkHcyWM-g	lV/l=W`|`zx?u\VC7a37.]IAhn{
B	Oq7BK3tEp%
]ZNWtut%KѮvhS+m5v=‡TM(Wf56|-D#v>`SJ^097l^4ao*R/UUHJk~qw"3ꆲZS?Nai58wx8oL=x2
|2V;Sr"yUc?zc(1WJi]_A[%ꝣq$2WWEY,fWJAz+(ePShݗyȮued7M d!TEdQdrFA^e'W2hi|ϯ^-y/MHm%dl9UP*UAWQ\A+WYay,9๮3+5''x:EKNi.:^KW8]\ɻBWֹTBtut5؃㾖"vut	ҕі.}C\Bg
]Z-NWr@OWCWp)U8M!\`q" N^vDc\cЪoJ]}=t%zyVǧ}d
WNhž'mPJj]Jtҡ8]!`'CW؎=	PJXuc"3tE(tut%'hTES]>s`mM'hvh5F4*e+4MhOӄROQEsNuȚu]+Dˏu.{7kz:RNZۥ=;]\Rm+BLOW'HW\
o;]5DWVR_DO,S{Wug
|b zcPMEk_ZBbvɵ9mAۀou2~D(Տr~4v-Fj?-ٳÊ#dC>D,9:Y#	ZrJH]%cx"NWRtutiCԛ;ݝ}r:]ial=]!J+#Ѧ4+#9Ѯ)섖+n(UFMZ;Ƹel2?Ȥ|SDyu38u#ܠ\O>&ToЅg8իւwhK:wջIQܨ.+9-!WڶmasmazZP#zJtu^pT6y> TVx|7|-#
O_Wn7s^]M(khV"_I4 -el:~c~43ME
7ݼ{ÓfxumÙ|t_ܹ3ijN͸q6ׇ53دIr \t>Sqo}Qxz?z1/ljwxcYrv2^.O~v_B~>-
u}<6CR8T{h1eě6m6@ׇ.J?tm޾YYo^Folb(c?+!ID9]lisLjhgIP@Mil[e(Vt9]M/Ϲ0gg@cIf-gSh6vzp'~ipsa'?=x++n)ѵH3l%n.Whӆ5羖#ۿbC.2ovgv0 YYr$;3[ԇ-)jI[vqYux*Sj7OIBK%[|s{+׶5=L~2<&zo?To^:s
r`~̟F+#0_a6|2?<ql]ƿzTWU~?_+V!Ź,L_4?7~iS)~q|ɺb֏p+'\4L(q$F`X
lJ-Z6V!Y?_;4;n$ZCzrb>m4mAN:7qr7?FȰP/_-jvo(rj~bsMX	CaG_I'~[Ca~z$+	|c /%iቇkkQPlFE7Gϧ_tc	y|ݤq~Yvݸvtl5j{ݼ~DUv72
#4{,||דۻyZ_BmVUX#~KϥAvv_ehuí_[79M{W56NCnpfw~cÝ/َy~}N&{[]mceGK?B';@=I`ا1CzU?6fm-vi?dPKܥZXzU\rNX𹭵Ǩš|=#ƵPC~൑޺	#6Z>o+@î;[̮I)=Pt~>V3Ϟ-le<[ّgc9Fv@پ4dNkS"JMELY:v9FT|΢F,@J.|
! 3JQ{aeיNm$Tcg3q;ytϞ;6uqY]f.F:"be\]>\/=t}d&
[tO@em$IycN	Ip)k #sɇlR'-9$e0LAMǶҙ8G R\XR6hV8)Zi8Eѹb@hA&^ &|5
12dXo$3/1KӤύ?PxX"Ţ=h@d+FSꗍƢ<sN-T2v]I%Hi+3sd`c==3M޶}'5ƶjoO0hP8PKٲN9/QcZv'uH!o(3`mFUyP11H)Ab.EZS4ztVAZ@|NV>'m||cuͿEzٸޗTk[OS:bQ=k4ZWS5ՔFҪޗ:Rġ4ڳF3֭W=OIąr|'̚[)rC%gv@Fꈃ'z%Gh-^Z'v@k`?JIlU}Kt`+;m=ViT>Ġ(yd!@-2w:9y+
{EIʚIdH82F1a(dymp~2v&a<;Zǩ#"8 \Z+)a3V%
#
Ze[04S؇%)"/p`7&kŁF3	S9	|I+:4;0"xlUbOδT\.L1'a!"{&BL l@ؐv`7@4AY0jx\
OJ;Nige7Uk&amݏ-c4L`elJ=ˤP_mzRA4lz>7=$߳TBs{iZ07;)YFQ|9OTBQh
t.Em|4>U7*ZrET4$=
бЙ8Gn2on߿'>U?qw>FשŽ_6>ޭOx}Ōj㈟Φ%`o?
PjXr iqG%fEf.2[YeD8"T]^00h(>
XŽ%w>Be'OIz	|Sr
P(b4Z9]¬UFƬ:9ㄩO`{uH?ϘG0y_kKP[>]z{jy6ή|fC$/$Olmdl"6JH1B"
LwpV!O%-j $(ob5
2C6D:/
	TzPp\0rz7ZXBdiШ| %J`cbg.)1X}gV=2nR=tA*:$7=tXI|_ī/}_Xa\"9JQd$2o9)<7Nzeq
ƷH>	nq"k{6lw3Tϋ2~8	r}#сC7[w|PtKS

,rje:#6n>tm8]Xj<&Qr«.Yw%kt
rVSin+?hJhrHΓF+EnN1
<8ٺ&u͎{,P|dIys!jK%ȣ:@dWQ~:i4䯦p~\l_Fyl|!}u
":L&
 Dia
uVդ}t>DBvb˰^lW^><2^o|z5|
~.~;_*wݻCyS+5stfg8H(E4+јZK#r.<\iSLs;c6Pl[>P5gd+~!3oiBFmHDɢG0H?[ҢJQ$mp)kD~qF'IũCR6T.!Y0ugY?A{֩O;jA=QfbJ7_uɣY=vP$w2}(~9f%
LLH)>HɒKƇi[-;(pjsb]:}LuVxic=@V*uwȐutIML0Tj|LŅmEe鹦RabmU$tҹBB%Tdlf,6dC֐N:#)>IΘbt$AZ>:1zH&	Bwo+k߿-bL1QۘwKw]y&9jj;~}ů과z5v1WY[Uͻy$?2Y[׃AB\mH#\D0?_l95v9i\Jώ@
,
%{V\2ftF_.gw74
oxkUɁgz^K!
lx-+mKN-A퉃I`A^j8HlA=UMH) j~N}gEw%[ߟIAq8ݧT[O#5gsѲ}s[Hy,ۋ
:>>>3
cxNzzmh"F,Wb?^9qx,aTq]ʘfo׳-%oJك΀|љ3?LznY튒7
L|%n %-	fXc3je7Y"ц83g1r>ѓetՍ2 ZmF(N}&i$7}1\$F=Qm¦J%וuqOq;~Wo^o?ޜ|	e	3
C/̿#?+"L6Yc=1~ZnRxMkV޼i6M6G=mUmA-Y?h.b&>Ҍ1!&lI>m](]Eߖ1[S`F9l@-8!:m\[8AH#uNHj]0.U!_pउ"gS]4#xR֭mN酣F 3HuZ1Uxk{NaXzo>cEmm\Ig]P:qFx9";Z
3x?M-xB"Uj]	AʦIBCTtQ]D";')ig`@%:OTv"
|Vx#+5%rֱ$gqt>cC"x4H1\Pa89*m=QFGE<(_W&&_h("J;wfƸk]uA]AL_@-s'y2Os(L_GW ~ʝEy_8(_Թ,+/~'oE©㣟h|yl_FH~M:uWG_|u*gR_/YEqtQz&	ǞVz.zdʙy;⏂k\w%>ȧE[DWwk%,g=ɋiN?<^v]A	Jy4@J* v\4L1Bu) @
H<hvh(Xj#VGSn5NK3J1^rD<-SpkݹoYπǣ45-ʞ,}uʮ
ʮ(6"lG0dB	/[nɖ"[f7
iBQdojun{"E(ψ	f}s7{%Ve
)!Oڦ|7$JV6ej~yu8b^
<ݾz!Iԏ6_DʤRto?XKo%-tv"!i]+S/k[o
!n|Ew@tS9"X)Ixd={n=;N3PCCcQ{
	̺AQlRO)BSg1/S\L`@Ou4($9!ؒL@p("e4Hmug /AڰHxRIoυC04;>6TF'
-@*YzALWwW%
˰iP'$/7XR͸Xg6=I(c{4΁$%2vX㇜tsow?K:?係kK_Y]Avy[
[z_o?^?^IWTNVqL	n,șޓl=ޓl=Y*@c+uH		 mD!Sphޖ^I2,.mWU=AlgXϵ殮44Zap#;VW}3dr{Oپhۉ0?o4[ͅV3gZWvj侓zŀet`Tɳe[pY;}Z.iړ`:џDoJJ8*/eAS5$%D.2̡<,e6#8譨*^e$4(jMT@,Z-	%n\YɨrK:=yz>S(p2Zl5oIU^=N)[ǩ(UpS\
p^N6wNi;m|6~?J
I`QӴm$@UIPo+M.9Ӑs)IkBEN%tC7 D\;%VȚPng4*F4NScpru.狋%QMn
8-ֺ[͛CEsars:WkJ.;
hr2gUwF:3T=;zt>-u5NlNn$厱f \=9nO5QƷ+|?]oos8;EAmR!Yln_V+iy!G77/,K.Ǽ͞
vw.Woxx`Qov5ß
^<>UyYN"wxF-^vD:AcÊUiㅳj9D"PvuA!AL+ugK=uQP4V%?B3!$ERHx$NhR&^egpd&"Rlb+;,4!y2@Ăxd=E:ui3N\̚iM`3p+W/LWecTˡ/3qYe4FfT'"@Fv+A"vq@;#G8+{޴N,FD
ʠz#1?oq؁3jFRH!t|>H-(MP(^;W%<*IQyhݕ77=[pgo>[8ųhr,2-SHkM'mZJ01Ax'-GU:9h{vtʳyhyB*ߝGΰBI]	A(`~}>2C6Qp]6Nrssipy+,ϛWٷeO=}7	,bCz
hQA@*M90Ht>N/6IL"98ǧRs?<Bu8y7[OV?eu(Vyuk
töCZ;BZ,эOO]4'vԕW3[IJO@}Jl坜=#k?2TlCELhЌ{xIĞ٠,,
B[h"]Kivܝۭ,BۻflmnOgcLM؇x{L%sQ^r%*ʓ@bpeX긶2(4HuGg)󆛫V_y{ݭ{Ev͎Ay8ƁHqNxF
]Ƕ?6=Rζx|/[v|\/z@HmZ5OA2'SUB(ÈJd|9~?d"(BT:N)}˄Zwn'>gKx={=Q2I.~}rE&7޷	#"mi 8d3,=,p૭,yr2+ZeYl)	ƉYM>E..傯>>-=G9qRK\*|JAkgQd̡Ŕ!;5MqW?e#q$G!{60ce
<†BT`&9l8&1a0*\0b8fǾQgf7YdGMJsZk-D9:X&PNa
wZR22CdQ4z&DL{1
?.$,.8?06F, b-"BaD{D\,%h$cA*Ijμgj[x9[[u
c\C_Wݷ,=mOyoۓLB9
Mh!CZq*&v$9r1QK8tJ2#N!v- "NqM|rA+ Bb!P
:tTSࠝ.nn@w:|iN46T#hrKK!>'d2 kTPSt[6)bET,qJ!BBSx&y衧zzڐ'EˈpCd!|`7"J@\z#=̿GJ(mjѡ!z)*m
b_*YB_Vz?W4D4&$xmh@&Eh,(K9٤qv~M./4QhO0RqGݨelr<"Ndws}Te
?bJsβӍ"86 7jz5k
Xu3OvU$VT/߬_j2;툍È093^ꪒDG/[Dlݤ
6뚮6:ڥ<+TAn;Y٬un'WѰsKM@N7Tv~ߗML&/?C2~W9:{P1r 囬|AI]=gܰhol^JmnŃ_.~j<;)X[Y]#Qk~v,6vݽb\NZRwܱH5
jK=<Ҵ˄͔1c=&z~o8ۚ#{m6TSvaB0m2V;ͯ_xnr?Ik
NF`Ԡ'DiMh
AY,CkT[oIْ̅>BH¿hQF)`Eo0D|@(Y4\R)•ƋAYUDQ(67=vY^X!w˫}ty¸>nq7-\1!J(o*\4؊.=0K'uJ\5QXAr3Xjɝi؄`,R%M@:s))\Zύ$)$0&	HJQ9?춘8[DB}R͵5[uu<7!scp|-_b4Z+Js\bڽlMl2=yb2 Ogd3oL%J$+#% 2"̳T=CE:%	E<
Pyeb2/C4E"Ajh
m|`
k1)tKϕzV}>TjvƼ\c(W5ڑ^dJCgh.ZRCj7B
Pv֨h(UXj#VGӜALh)c+nQ'a
g}z@{gb=e2.AR!b$y‰6*xojA1Nrde((uIwx,*D&
)J',&Άr͑_21W6L$C4'u>~8/ewtTm2_[5+\|UnrZ^FT#d>*K%.I;(Hz)	62IwP˜h@E$S'/@3 !"i#GBB)JNˬ ί{!cP8q8cPEv"K+F5Q=xƁϭ
Qp1"&IeLz鹖"!hUEX\VSR$8,s	Z
%4^*6wb}4(%WB$m$J)Bm9gAY d"d_eRn~}vgyY1j!x2k^4[~h(kf=^~>RAWY4pfN&8hLFȑᒑ&2Gj8
] NćCr<#SQB9aG#[Q}[+>*yYNdpH
x:/Ök췡?mgyFՇe}bC=٥;D3l5(trr<<^P*ЎloT
\ūQz$q3|
j	0Nmxb_O5\ޮ{08goW6ǫYs߲sGqZՒ-	yK]Ͱf`emfuG
q|_?ߟROwop,z\Kʥ1K,lތ~{3
k-|pghJha<8$<45zך^kf;/uyf߸u
R`I6 V:!:m\[8ATF(z#n]F}~=2Q	Tp&A0zT]{k&rSx02NCVlNl%]2JزG/lO9,N;/}U/!IἸbP~PIuJʳY*(Dպ*lVN	);.[\<_^Fčt11!) sf	&NRX'E,Z*RmZp_с/|*X]aTL
"}1\|^G %~~Gǃ	{OT̋}1 Q-=;z!v$BrO]2b1> FULk^	K\cI )S(|.J$gDdN4.P0"0m$@
X%u>gPwQh(܎uon} {r/Ă_{BgmOP4w|\fd$"'l=_x|j~o-y*|gW	"'1^LOgǟi|?cwOȎ]ܚzt2f^4)Q&ѳf]%Β߆:sWjI~.m}ͻ)杼?-<%ْM
lhNmۯ3Eph`1P1~gv5s=o`~_j~3aYejr:4v-WwgՆCHvFf/WWljKeqR
_pWm>4
\C,-WYJ{xg
`,WYZE z1]lq>noSjOO޵5q#I6	uUvOJ*3He+&q!	JU 0b`:K.U&i	Ƙcl2cyqNG0S,V}TT](nq]vd`/~Lz)IFh}o?9sèj3~~qLzR[0EEJ?lw1y̮Uu3{`Y10
.f`r,:FYfozd%Z@
F/pJÃ&#V5$%qx4IZuIJm<}
WZK+XـIJz)p=!pRj,;
Jh%M]α^K$-gWIJ;
Jj/*Sj.x+KɥU}gB*J1Nȓ*,I<Fޏ?tիv~2~6_?
`su/_fyUrC6C8x7F@Vֆn76ZUM7Jq8l*/#Sy?.#AI_0Lq3qtuOS/LAX.rF岳Vz)Yy`҅qc$GY~AQeapJ4̉ϛ@`B<UXj;cƌL"e4zl5ͭ2%kzi8<
pdj?g-:\pMpW$awV.LevG*ܾxV	oߵrhӑ\ؤkۻN~Pkq1ehNh]?.ۻ9tfG6e-Z[w>yxwբ畖a2mo1yr/)}莎k,&~#F%B5ß7myn9S$*{~	W6d!Yܺ02È=3kDV¶<_f*o>THu&;QXkEᕏTPDH]lO3lq&RsGabI͹,	\N
!Q&P0Tx^>2(T*uZ_q"GεuLz%	baʼnaoL"VjQ#]qkC}\Jq$\@K+ƥ2VXEL!g{zWÛ޿*]
C_Mk=hkyoϺZi~JSrMgaF}{߇aoe^oix%}pyP$9ӑKp3"ا5
k%0l|!/8
`$^xi4	@m&aIY(z'Oɇ?y^Fx^tR	Z:A]`1sT#YZ^inکxov᷍є>f_mnQ3k,vo;C贩9:*0b]|LL!]9	IUp;oB!gh՞6#pHbaNʢa<ոW	3{X{t=[__' fX\"p׶ZIdKESx~\#O\_~ҊEP쮑<Zk%9 oL(O4='ǓC[-.tȥ`F&7d_-bMUF6ޓ\\m|F~1C"h9!iZhXPcRRRq2qв6O?W5{}@wWrpʰQγj`)} b<*
88`$p*roDe6l6pSJ$0xȱYdm: 'H 	 Q>rd xMxX:	ǡ(3#Cĭ5̬ӂkl"Ƽ4ݽ1̀tA=8u|̊Zad(QHLd$&:łKsI&tW$3"f.Nڸi3):XJ,i
iq$2.mtD&˴G5:\<.
O̥jG<<\	 lt5rnEmaʕ4a.mH$Owrŗoð[^Р駴E8A{BRGb`2*dT\!W>pϰ22k$
X-VXR_oF̡3
m~
L޽mm!Qh:fDmQ\F"wYaE=U1X-XgNST"ˊP`+;<z:ym!(pzG
>EpL1L&#`Y&?m;ɂ<7Hz(d<FzF3 {qAHz<1!XɍwR
KPg٬~}Sm?rDv	)naz{/du]t>+qs%6:HrJc;.$R@tX\Ldw0T7siz9:b[htec]۔o)
/G|0{
G8>4MzQ"7_RϚ貵	e
z*o,FbeM=e^}3+\'Ji9/8]*tO+oy1S=O,B틭ƬnVZ
;\os`<פnZ7BկfF6"c{oP!a(V)i[kt\	`pH~a	btFJ1L6euVY+QwژEF&smv[͜-!t{ɿ),-KӶ7}jBcGu$cD1_ei|<)fJD
;+O"\ZVh[Z[V@?9-%-2hA|K:LE_,z6Avz>-%x'RwlvoG\7}SgZe^+:8s#˄Nt;&uM:n4
ůڷ-|NO1j{׏37ϻ<,93Նh^O|>_wɺP?勒Oߜgz%wgr'3؉G7oc''OArΞfkN?J|Ӹ,Le]|Hߍnb~aq0/{Aa1tZБ4B^un֟t|,-|::l}4t/;?X$w
,^h׃'i@o?5~'jJOC.7VK;,zcs1>,M%;ӛz_=MZV_ZPBp϶]J!ZikU)c-jsׯJ=ޚ8Pn6l"ЕuQ*NwpfGx2_2֬n)J,\)ˑIfN1zzeBɔP.>L(knG$tgI*CN`]T9`	8Cq<'R̹rزo_7Z[!ŒO"X&{ꐖ,59MK'"IP"m
&@DԞ; e"Kjc\(YZT,\A_r.v_/Fa¨h!BUNA$B9Y	E-ݱ[aܗǮCQǝ$`F
Xq*'X\p/6X]j#:1A:0[!$O܈O/FqjjpPQ^UD\10e/B&&{R݁րՂcFJzdTBc1vdAB"oW|<)"4Y WS1_?GZ-6~QE/5_BbzúʢZ_9h~N
	6τsr?ބԯI6f.ب=n(Lp%{;2tFx[k5>E*GGYdRGDf,l6ugnv-ѻ҃p4#~|Rm*Ew ҝѽP`v~V	ǹ"?Fx?50.?G=lxvn׶*.nr]oTVec4}7\Wi=F҄8K,Lʷ|?J?OcXht;eU9.3aGԤW^_?6q~=z

鿼M[VijoѴ>Mw=iԼ5z%ijAoFt#Aߕ{ӷGWė߸V{u+r}"=	u!)/}̒Smp6,Хϧm\_rqC F
GMrGц\J_*E3"i#\f
:U6uJ԰ZHP([Spy/w!n`p	LLKY\y,Wh*[mbЖ6h/?ivB6Q(M|guJ|TL늧IhBB^rD)c)a
LFo3XUySGj$Y2y(A\&DsC,*jܣ5@&*-eOr9񉳾VרSQt3Irdni59ؑ=;:~?+gSz.Ĉjx*:
ke<mA,OXPnDDTQK{4'@sčsGXD	''tF(_GmD%Z{KLE|?L,Ip>D}O}:PlPE_J"va2Kf[F>vKИڶLA H*zU|Nhi'X*z+T|A\:( $ΉsH"Q*1(_
KXoGNdܢә쒐| lyäeLsձ24\2]9|{HǓ2/XJ?5oNv6+!(s\X>_)vH76rwtczgtc]71ߠn:CpE;v
\ILmbdI2B	W%UgfޕAjMvLrԿ0A|ap?˫x֟i}w|8$KH+߷!.݁b.]baJ;KreebO'>3(o0<
̎YBhdldOK(&e{]8OWo0tv>i)D-#*&.'wSpL50Guh0xW?.}MS>7
w]aWo&l)܈GohDLǧl\[A]:*8:QYEr"sm,@D0hG!x
_,cm"*^_KzKPΣV(!&FktI)8JHc:gDY&s%	/\Dܤ^9rԡ|h*RC#53:LiNH	&5oC{=gݫͯhC37?q_&$YPS#IOPGsP5fU.c–]z%*w0c[vLThpO~OSO-iG.0mFm-u+e0Pq5FP۝ ֕}ۓ-Q1HT,vbt&{LA,>!3d$`PQ0SB5da
;CB>ǕB.2Kə̉hrw;|jSVޢy݈]ALNw')!.e)X*&L4[J)9UT<bƉ"5"B~^~K"GT!ܐ*`N-sne-Icu"HXl$HxukႄeM.:5^'g]KXï_r~+gs<]t5ߕ|&'ujZ(m4E4)]2Hl1asrU2G8
aC]
)XY&(eYR:ƒhR<&ZzAL Us72*հZ2{ff?-->ƹ?㏧O\Ĉm	91L|Ń3R<AD/)zIJ!EMM$Є
%r]r3#k!2nQZq1mAmP{`x(7{m Pg'
cJkwBa=%96yWQKD[x1X(=!(Q!IȢ(lDT	c>ɲh9wa/b
0n"V"f@oip<$+8#MeWDL	P`[)%Cɚ"I3"AI6\&D.8RVF͍D#yB
WIe=^u \Mkzɦh\pIbLI)Y"BD]Eܨr##7Q.Kl2ƀ[k
p5
nX lF`b:óns5p'YNܮ_ׇQtA'1G?+,?{Wq_l'Y`U,++*֜$DPO rA\R2>f㜵NM9:*FkTf<ǭ/5p
zUCڃ<"+bdW:IԄx2&֣,J텰k_,UtBXYL)#"b1h#P*tZCks
XL4a(Gi쳯,Nٹ*X
Stxz~4;W?,@mtP(ٙam5	QQ%ZP)T2RxL6uѶ
gn}Ud<qF㨻!H~Nq$x_&ƭ76Qo2'mM^oE`oecz.mSR>m~9~f893-KzKr暮_lp:UvOOW§Mp,yxRfJ R%:z^'Ї,W.{b+\j.䝯]KWh0xNu󆍙9*]lgu͌*lC>oJkL+PwZ5񔪸eaV=1Πߌytd@k՚jݶӮaIMhCLۚ2uRb*}Zh8kƕ!r3J "^hmTљ$1I̅F~~.4-#pNOx 
%$IE	HBɑD LHДc0cS	V^@OY_N%,58½UCL(J)v(i:LX
!@㑃MOb<*X-,`ٺyA%gE"X6/eHdeFK׏Z&0Oc3in39%FMR+S\M
co,!ibKn2D괩̺a&^z/U+,0*Wyv
DS>nש9q6gr?GuN`#B%~Rޝ,iVX_GZZ"B6j$JGǰ1c=ODl0.m@O/fD
eo΃*óozYa1?Ζ6gn
i峚62)Ӛ_˾v^IbHSWA0N;Χ}.^AQT
*5yqޘ@V1z&a	8rS*p9
nyD=nKw'#7KN	@bfQYq
z}&i,2Jo+t\f
8KWAR8m0
gwmKŶg#Ibz]Yu%Z|Edeeei~QӬL_Dt4+1^g⦢ˀt1t0G=#^=czM	tfxܛfCVCWmбZʄ2a^!X&HSE57V"FMsFV;|t|$9prs
$_zc`gm~kHUrp׼TxTRa H+i\lV`v1qÌ#P*uw/r~4;ؙV]Im)r1cXJʄwCQC^R՚T.j$\.%6!tG2	ǨU1k`:iA@p[npԷ6
u7Bn>-1L=l|9uAߟy*\!M2'grGs	MN=*82<ZA)FR	)MZz*0K_u wtft<]fHFǤ"9%âҌ;$1
m?ϮF;h(~AQ*U2P(,DK.-@-H',w0%H&#H[D1pNcQP032B8om*s~Bgcp-nϒw:0a"q+qi8vflsǰOqYѤ|1îRp.i-"Mg݊'F)''?8/@wي"Ua@2L!SaNқީbzf6vx՛A^Pճ_%14fWCR+&+ط;)첣ӓDAaT:an|f`~/tpML!Dk{QVWHb_(gDc@)usm
4{\"ͮ4|s,\èwٍ5Uy21¯ɛYe`vdt1i0YT_풑(G?\+ Ci$Gb.!%:_fX0	v ax47zv=b0_8U^dӨMkUbY\59Sl$6qy?;:@
F
&g@_8:y7Ϸ/O/oN^~{:yo^‰<88zM|1B,,VFO/G̥@:$0ɧkqv|P
fhymceҌ-'pk迿aK6ꡩb\uᛌ+\rø[Wqy-J܅/ɻt3[ȑf|P\@E迥!Z&xU1m4+•*}DKHVi뙡EG˅*0zoc$nmK^h__oTJÌqQzm6pd[ӁX1$^H#iRgb{[b{1khōrʢ'_z%A\l~)_ v9؂CbR9cf3qx4?w+.[Jj6vRa+kܪqQkEaR08#aS#(\:J4BI:"W84AH
pTKF<	c	(Xe[IMgsZ S2V5R2}؄"YnݲcwءBGcv@iTL<}.cOfվӧY
rdNYn Ag?k#$ErÜ9Q [cI"Ȍ,ӝ
?ǟZOyK	RbOKZ*W3؁NS{&MIUǏf$TZ&=oHDQ3tCt
8/@:Iy	U=ԚA>
d:()
Cb=Q`͍rCD`m!Q!
Qb١VP H'T	dmxjޝN}N1N;MYWG1	b
k?[qgpTl+W~'c3:瀜s5(Eb~v𳃟{
?
4Im0/$=e
IyY%Rh~,3O>^_XnE];t\6UlU4K8SG:1JTNDK!zaX#M=2RV9ịUh4BkM̨ʃ	Q9b0XJBSELt;qgx:q]лuB\Wt3iR)վJR)YA> ɴStKa&\xs&
v4Z&4QNQeUFznUfiߘ6Uar>(t2uW;s(.vUNY)Py`9pj׀
z}ML&YU~9?·Lw3~p[W
yW3t!0B!js&\1nrix;={~{w
V;C˂rj䉱r-1{1G1P&c̴z!cFY;
$K(ȖŸ"%М-=6
KO
bۆEVֳh;C?v\H)zb4n0ڸ0׮p5U ;wvV
THϐK,N$ϙw"֊+@
hU-{pK ﳫ4;Bi#I&>]YY2㝨0Q6oV &P -#:nC#e*LO!D#*/t(joFI*S|6^Ԗi d&)IIdJ	KeN+7^<1hQ@{ɱ@kFvJÃ&̔@RfCET<ڧ~.m=2V]ŠOT+x~
I_z9?lRexPSWhW@pJgUu0
oW͒kXS}]Ľj$c	2&k/?˝uPKt\:FGx,@@@//ä5B`RGTS\6>)DcnůNVWaYemeY]!]+Nt-i=\b_z?IEaudS+g\g rN%yol\ΤyeHLdǠFjF8Z%>!D!e!x0kSIFMFS4OHKDfltbK-ٟnOUmV7)m7I|ZR$CQ-lI[F>[ݻ'M0kS:EмƝD6wzg bxUu
a:-.l]ͳt-khYvkn7Mo|CK-7"=eOt$\܊m<"RSK+y2~p BTKAGB?8M1
tvqUnW/TǥWgRգ֘<3\=N\J'@WR+Rv]zLwU
;j/7&\&e2}75o4uz˞W.fւF\l6;BV7W|	&5w! X]l[>on6X3tmu$XLjAϧvYA/SEǸ/gq4]LpQݟZIVhroY-i,dLU|6.d]iV/=I>7[ّQ{
[+oFȧ_w9eПZ(kC(FGTK;1TyFc!cM{3>~$2
j|@'	f-]8ffV(]	__U?3DØ3Rg?Et[a1~{V{W z?K י!:smuQ=|1[a,kH^2E.*?^lU(K}bzS
5O;M-]Ӧ]YBs.piN;h)8&n[bɛ#Nܹ"tyRr
pE}}=Ffn#sB0Wlfw\ofB+~BJN&q8IoLRۻr;$X:mY%0Kn;K;NѾP:%J'ZiΫ%LӿXNinv[<;s`X;tf\{mYzZs0Μ]Kldʰ=лog.p[YNc
G{C04t@\עsW๨kw}k~%*]]a@ɐ0f@2xݜcJW,Nfr4K&7ɻOzLnM9.{AjRM:$F'sڈN**4II櫼
`d=+o$gm`0~oTjB[a&78ӏn%!2;KOQ2J/KR:/KR:~92WWw/KR:/KAR:/RJǗ֥t|)_JǗEJR:^J-nb.KRb.ڋRQ
b._I\AO%-I>Wr)ֱ@7)B/bt|)_JǗt|)_JǗt|)_JǗt]%
%-O?l=DhVHCkKFDŸX]2v2ܪ{dS0,DW:'i.`-R^dKCy&HlP)" ,Rg2ه}q X{))qJ7Dd%0llfgy
8J/6iT:qPۑQ
n
5Z
o>h|ԼmL>xX"'1E7S*	3Hi4Fg	|k2"x
RhfQ{;&%GFRr-VLFV?{,~[go$!ԒFⷕZiEniXoWAxH\!~Z*c5LM=l9Xp\cT
C`r1Ԅ2~ǭ0Rv(⼿שt1^cÍu>yt
&&VQR1qV{t#iZ+}qpǽO9~Ĩ{9{g2|ti[NBqj^ƽ)0tk}s%KWh\
JX05P^w츽:6;rD);	>
;0!dꝎX.wP1`o	X1D$(E#lШ01O5FjM# 8e(Y50$5AZX,@vGE?y]`DnP{+%R.fgTѪjd3:{+F[#xZG_¦BT8X;ly hrDT:{Q6/JF(xy#MmXpQ[M*Ffdj0DȘM)qlXmg슅<3{/ٟ~ч||;v޸}
j
#z"ÕJY@TAh%JrHKLޢiHƞQk	%	6`C*)	Ӂi]dΌَvK&池vq,ji}[`qX1#V)m4!(,b
ƫ0jC"xY1!(
	D90
>0G,*8َQ?Kx,m슈23"ʂ7H\c6=a,0z J;]*(+9)\	6QzHL8cu4(&LR#1s-w\tp`-x||Lwvɮ
.\|R,b)1V@HcC(8&זq	fC`&2Ġ1e(x\<kw숇Yp:sZN5T	x?>#C ~25=)YM;xM.YKNB[*Sjk0x0wmdgm_*K=Ѹv>FX$Wz,7+7xj[&6?VW!w7t,/4%*O*M쵲l)++`5ɦؙc&R)ZNiYK|DR=0a/rZ{	"3/[oh슱:9-[L:*F^km	y:;d=ן½
~G?"J X
Q03-SAjb=L!Ϣ^kH^x>WRbe}v襦p4DL `
Xy$RD3tgc`W9)ӎs]iưyW|nm96=w_|?ݻ/$\L24"(Aabiń-o+8ai/϶qTFe,WE03L	='s`[^ovǾo%w:_\Z$ۋXKj4*OHrVx!vXQOUqƙtHŲ"j+ +LR5
|U
>~@U
Tv]_`tFhon')?nRnҴd7SV&ϚUu?٥{١{hgnZޕ]y9ߪ_:Km]MtЫи~y3865mhni
M7>LOdmI^.Uqqj|8}YjN1E2DIz")HP"me`g.f]Cڦi7xCG/
}R*^6W/q\6JEsB@*w†Jf3{Ar7y49^	VQu+
PE?b0Y=Pb>^lr=tZ[
R?
3mX*o:͈1qǢBU%HnX~ţq]Ü曀hCLۚ2V	‡}Zh8kƕ!r3J),
ALEN0ڨ23Ic(̅F~z.4-#p0.x 
%$IE	pBɑD#dIДc0cS1V^@/"LnƇJ582
ʅϫ9
]fIQJovn)
Ո}[`M$t	&b_sW{^"l)(#~]$ۛԂNrt#oc	e& ¼&BP5;jUTsc%bDh$Iкܶ:E-?b<]30xW{ׁ{*r)pi%Y#ӇlKkP[@/643N5X6-P.91GJwCQe^RT:ה?:$\.%6!WC2	ǨU1kp>%HKȷ]W5p64q̸xBVV/Aw9)U2$X,־i7TeGsάM|<@B e4R
K%6"UP8a9\r7)@\:&-Y)HHitL*"QAlXTq19`U'
L<|=KĽPIyK8znb(yBlNY%RzFxjRMcF!{,Bcς){ww4=Ѵ甆T0#$e"QX\Z
 H,Ar0]Bڪ ps* ykSoE[:Ejw+Wr4Yq6?8˾	j^#gɄ~Ӡ)0#>U6.gJr9~]
;!d^,B=M9Ae,|~PgUV0e"2
d?FB¤7F)(LQ+;`W0ylS	>X,՟O:$UX
n}w^^/ᤌl|]'6-m{9BggG9#czMi;{nV0{Sø%n'o7pN\wbTlaNujnDSm`.0seqY-wk;gɆ75lqa!4Ǎ#1zaH0rϗGw@>G%
Ӆ._'k&'jN'4j\ȍl6jÔF矣MQ8=KsSS8KYX=\|	7~&}?՛w瘨_>8_
LC4/g ZFxa93Jgj.9Na˛taw@U駓0g~JVV~ƅ;/O/
-77|hƛMdhd\;V|By. i{p]KZhQxD>s)(a(fFO:3ܚ*m=3ԸHr
>H%/C4xxB%aFm8(68ec2@MʩZ/$	rsPa]Va]䌩6úfm宸΁J.&
WUx
hmg)t+[_Lb~38m]|9]B7frY٨2Kҭm2tUULN葎{yg<)S>aR^X8NYX4*kc&PfTe
ۨQ1,%!	jCe
n
)x	h4MgGK_.Wj(fL%BkKU*Z4L{	sy*R a	,^s͹I*M-TzA~p9lFduYMY+	i藺|Kw^^zxVXX?bϩub=Gw~k磡ɥOWRAh{h󽒏d)Ӕ%RT\&[.s}˰M,qtAnR-_Er~ݨ`tU=wͽ b.x\*
,??b]Tϲ{-T5;F&ٽ|N?MO˔(ֳ)
HfT;ӏJ,:£[ZBeZY([臘^(+EOҧ9WrdG1JT\K_W%*ᄙ\fxT.L~O-Ei5/XFT>%z.>Ԝᗌ5GS2&鱔IJz%c~]ɘϦdcp]%p	?v76ή>lؕd;HsQecs]$%']2ty~s
hș1WB
[F.D{z0Hܓ̪t\oD;4bЖJF0@XkQn&iHܑq4Ncwŝĝ9;Bk>TRԞbq؁CyW}(=j
."ƹH)gycDMbTX"Sȳ( #<)pm0A߮wyMۭBzXyƊeBIe!obiK@%(d,F
p@%[	8^8B;1g`TFe,+J3$H'sT*rU
G[o?{NomJq6^Ʒ)Պ}qFɑANiA΀4EXsiNRxRRtf}n23uQg_EGWU$GbtU.NIܖ\_hE)I/i]T>/X#
+G-*YFJߗp5hMsmru_{-}w4hjf%</lvo2'-Mmw9"}0yإެuHԷ;o#0t3~O/nEѫj:.#t¹TY@)†Ҩ
/g2bnMn5e\慴aL4/7hV:d{::$SеJT-sԃ⶘+_\
R&'Şa:]*7Z
\lN`mPp a?akn0z\͈1BU]$Uu`n4U.;-D%3w
+WR8ȥ'(QM419@C/Hg[΢ъa80yt
Ugxh?O^oxbH~0.z7qpAo/1~TKZQLR/o!JH4Ք(${zOk):9AHh"`$p@sUJ7"14S]]>:r~鸚l2MC_\\kO)R=Vs=uZϛH>ˤ4A]
]>=x\}nj5}ՠV~BSip'hujxWj*5܍:]rVA/W!9Ts#*9Lى  rR|0LI_ʃQࢷߏ$\\vo٤.~ ]KfL߫0/8p0{6N>5
g5䡱LF'Fs&Rܣlf+ً~?g1KN<1VMz0{4IY2OrOWO&N;[(]崴-TG'-!Fmgs`H
:uhm?yL{ۧ+,Չ;\2->ݝ{Mz؎?
uf9:_gj[J6Eun
Sz,7ڍ34~E`ڸOs@N&>6$VX%v
wZ`O{rx%;Ky?Zˆzl Fg8CASn̍Ab
!!EJ!LD4;m5H"ں5iH7FP
(΄UQs#HPFzU,*#blqU2>:Iɶh{\qYbLI)ِ ["BD>2Q.(06FFj2m,dBw{kIǖx
n|Hyv|*Ll&~~T‚Yxrxv9+MB0qI#81 %FX+Yha~YPx}x}xK1DR5yI9Oo銒5)G<&ԅ*k 
FwژE
#Vw'wQ(3H>"ޗO&s\[t"S>;W>i4=l_ȸj.R*PbD3,F)ikt\2aOuv!;򨖍R*\{J3>NkMAgUX !tslN*MLksϟR<}=!Fb=9|(!ZQE2uLC_Ǔ44"BkŤDK|WϾWI	/Ɠuwoc/1;ضnu GIm%[wf;dC*Wkrd%N1zz+h4;%w\3ǀ%Xx{DB524pJ
S::ghKugCj{*XV՚䌔qj7J^M"v'n]Y'MBՇÎrB-:[nKx$A4dQ{Z@!3,qhYZT,?楣_u:Re,K3*^)
Q2Ge0*lh!*[ B9>ZBFEv>E{#o[R^MV`/=Y7fp&`5
wQzdޘ DCˌSK^)k-^_Dף/Mϵ4m2TQ^HEB\10e/B&&%]cA.rOiZq RYI=Jh:J_A!CWX.'s.b	>mGq6RKCWߚvh^L',8*[K3G~pG9vrh؁ XWA \
!0 
*78JmiLˎpu-) /D%;81ZJ.たN"3fqnjz}*GGYI@:&2cjĪ,q J9VcױOFSk:Q,x4^ͳglU[]>dW]j7Ub)
ϟƓXn}58
l>-˛C7$l4?=g$ 8:~,?saᄃ43nsjgqIaoߏgX+r)XoTy4J
=?җ.lWgX7wD/BA"PGǞ;"Efn.\G>NF5R"@'&6A6LA {HzS"]V#Hz/R"9Bt,A	ǔq."2!eH!2[e/\FN-q.rNgꗄ[["%h^;vݱZـ"qyގ6SD\{;võO8u8LJt-J<'N}ZiYλM`n/ILW_J|32!&meЛ{INp'VƕDekm#GElۼ/0,0=`_&0x)%$9~mll)Il5E٬b(2-œ'Ůh~ij4;4߯cٛCwN4g{/Ki	\씀cj+kZlj'o V{_nKwdk˧$䡑I4o׃ΐl~-{\Xh8.-z!y:$?(X֞PeeR8F	D[8rʑ5Kɇ,L
%/
i2Zi73^Px"5W zD
YXf1eHчP!v&s9x'=WxD$Be6#KHѢ(yG
Ag57MxBVSGS+LJP6).L+N'uVI
Ƒ<3m&K}LGdf-H
SrALˢ?s>ҁ~z9w8v&ǥigA\PF_˛%}MӧaJ4|K	e<'Ċ
i" 
EtLr`AዧK%h촔-eRt:>X4&4Ib
(0SXc*[(ՉP"Tۨ-Ya&BBe2N*3/^^J61>30Icly]|@hEoQ<]O!v4S²S?.e胟P76l&V`!9'i۬2@FvļYb/ǁo s!zf! jIrdib٬KqWDž5¾HWyZg	aO=A_
ۏ+Lv	84WؓFlO'Ҽ++	
i{(駌.@
iy^:E!=h>&v`a8uԊwO.=_y>oخa2*17*KSR3K<	VІpЇʉHDQ,^;hXgPׄШgaJ=^("5\sUbG6+mw,m}Kj{We˜=%c"
,:4K!$Y%`UpQYZ%+Դgsz1mTK0ZdG//|%FB2i./@gXNt23kڲ?<ؽ_OW;5O<*iFADI&>$H2a2Vj"KmnNG1_gp	=HdXaT4BU%⒜9ه`QK(`
x3 '"K+6ӟLu{MjPA:1A:0[!$O܈r;_dɗ-RR嵑Hȝk1&&ED z!4""5HqւLnjR+zc
!8H;D]!c|p'X{`cZ9	k\)2W}/gpegg)I
w쌺!OiO?ٷs%7
Yw.}7ћ.:A;q6R}Ntu&xL}XP|>xN@q0LɯFKe:qI.0j||Ӥtg\E|]N>8_fK=y_uկ㱕n>ΑNӨN)qNo6}[8iiK
t078Yޓ&jD#XFt6nFk:7uUַ:VWnTVŽS)
MOCrS~;V
*4'~Wv?ī%M?9xwʿ?~w/;x7F`djY/3[XL?M7}{5;K]Drq9>h,.33Ft8qLX?Ccs⎟megjA־1MYw9:(;=˞|Va;/Co^%Mu
QْJyk+T`
ZF}fK^`xRѳӒܳ/UAGWvIhAB^0R8R42fI{KblI
!4}^.ʾW'+J&:64=H_nrc aظ8БZLai\.\6lt%,TQLĂr-$sA"xnHR`.AHFfD7y8$J4Isa[{fyuxe^ܶ}>D	v5kAă>;t&!*&tqW3ϿLq
.j*O׏4Xq>~eGZ'pqm~k>\~3734dc*|[wySY$+_dg6=®R= ^*:m{@|g
=l3QGBL-{Ѥz-
nW	^7xVuhCͣtjzAx#+0@oT_@ե9ׅ0c#{IAek$#)=HmTos!eӻ=z|xmz<,{Sd~>+9`mnzcyS7LΜSp
|twm^2>v5/egn#7%nn(uSI|fsb&::CQwI9)(XXAu+HxZ*AsL2$,RHh "LyJovxb'2{7V+gE$铴0kU1Nl8)cbNnzLMWqP{E5M[p~`|G?ݶTDtF7kAdR( LA*4DG&.dnwRe4k"uorj`Rؽ'ۑ߼^F,S1uG)]'wՀBlIS:EM+ʪJ`$:`yg6"9mf;¡FcbS@:]
DE# :`vW?\I{9yŒmC5{Yw{ׇZ}mmV{%Snf7E:7"gg~4Qc
`;0Z@F8aLyTrg[BbfĉRc1`Oi+	dF}yx]Lߎ3ߞ[HILI*y@Xu0}d@\#kVr7%Kұ/d2I$qC佀ܺ{WTZKWYJ.?)^"rGS#zͯ{=`y/훪rP[Tб0,-2f(
zVT
R[Ɓ/%%Qwaѵȋ\˜x,!)PI 
TlfXL21ϊ{	/29\<їz}We4t::|Fl>aQ\ra{%Z "jdA>iض8fkd/:%UM|fp
* L \TuV9i>{s(L:vDт=$oμ2cF%R^AdH0*Bjlu>Vfl%)k5H";E
cÑ1?	EF?͆sפ~sVq(L>1"GDh\*8$pY@Eƛ͆sO"BzQR
g:Cc:I3C,YWvC?M.]JeXWdӷi0˗(f0Cf*DY$)&Ͷˬ%&g,Pey+#0
}(7"""E#!Dɦ$m -0	f	Ӑ;),ْO@=%-QMˤz!'Γ4j($/mdl+sPLVd#bG酞yJXBMe&X?}Q$Fi<\yZA><WF7FقEee!
hpkT<f6He1g]RbfZ>*}=:{pбAQԚMJ>FR
mwaB uNːVlBQ
&k&U.=Ɵ9ɑV1"A"6Gg)yd&=9 O5'WˉjAet7?xobe1|Q|iu旓Y/>byYbڟtLٞh.Dfkgz׵+kB}_y~}[	O~4N1'o.ɽ/8=G[߫#9l&~ 8Z7 i[/&&ō`e6Y-<^̒ͦg–wy=Rsug~ҝ\/&7[U{nyH(]R:	Qq?nNЪVa[_]vcqQB2sKPI,i`%Zrl!O#K7bN 0甔oڑ0Z
9ym;$YJ<)o;u)%aPt?9b


 lU笟>kfKHCF0Z&(E6m)K@*(YArAYH^%CDَ! 3JQ@Xu&Y`#wۛ
o/Zrq70Wxf=:	9$,_z)to>8>L:hL)YG(Xz֑\2桞tV&/72M PϳftM!im!)<9,Nl8'2&Ȩ)gjmHxD2IJ[x)ZHㅆead2&B,Tnb'`$+%iR	P#L[Z8a,)أ
4Q>[ւ-t,:ހJ9}PI'WѕTíQѱ풬ul@:\}jo];6`nIh_Ulہ2vnRpf'ν!*8cL6jq=-
6z)uWh]t@EKю~Ol{;Y+-sp]ko#+|]ܶ~,&{{lQF#)<߷-ɒdec)v5YuxY>ayy;@sJ>K,h`[ґ B2*Z驎7;mQ+slzU9jk@3Ƿ0ͱ]^?"G=zo 8߽ηM}37?ލG-Yΰг_;cc]+R5rLTcXݺv(*P*KWe%6WWBD\Ű`u1UK10!tǣ^SVzsu,cY;YvH
~όt/z"8Xqka.)Br$*	+AU&P't SG*O;OE=,M)*+"lL`N[G%4y,
@1P༷T,V'S.ygCgp8O1q]PCOf&z'wgkzϖ֒cH0$ͪ3>܁qz
x2\$κsHw&պ!VњdYj2ڝ$MD12۞#mz+u4o9n[=_7.|pZ0;Qi]:?"Ko7_u^#g_\{-Q+\>-L,yc{|̭AR>.ȹ6xሆHyrTYW=7NA4Vzo]S]vx(ijG%zs7;77Xoȿ1r4-b6ҳ[~TMWQe"!muߩL t*R%VI`^u[}WGzq^\VyY
("M(hD$yq/ kI	w0{}j~e"f^m\?вŰs=V=ճ`^\l^h4~_O2MڐU,7y*W"pp(rɼoxAdQn{$1o~.CD_RsZ2E8JVW0k;ש
qVd5yoݫ,~P" ӄ._]h6k^{3ȫa˒Dhw\ޢek:/[֐rW{
"wWѬ>ޠwWSzf=jcj53F\$_}9x)å#Gź(֡*1^2CɅ
WA[2SeAP-cˊ]Th=T'ѓ
J<p,hΈ( KJH%x4ve[ :{~UN'j!r:8Fd`OP!!ȭIxH
))e˧@ʈU*dd8.i-"@x4TyĄ†ZL݆
ٍEf)R+'ϽZrV8TjsYЁ\m^&r>UMIтQGhe)2:Ř'yݣ8J	F꘬gA4)	!*T
J#c1qHu]bq(X(:,<*n_ef\,̇oTn8fFkU1rRHW4A]"h	J$ό)О+	8#Ð=+\逰[NMLXV̅8ۏGq<.ڲCnx79GeGMJXg8$ert^MtLp'D-4$vIzEQ4bMfB8I}ٍSx*XL?ED]u"n-tHK}4tD䈠tI΂{SByB9oO!|xGe%/Al;wD9M2Mb04@
$[6)bf
˫dRPP;)pA QAFi*+@:P[AL]fn#dzzIp j)#4XQ\ȌѢ/-ekj}ZyF:/f/;o1Mjb!]B݋~BY/lFzW/:BX crYEDIUrG	B$,$:AFM-#X_'h"j:J&,[CD09ᔾ0O&9g
iX_z8\#sr|>=Lg|Qi/Ϙ2י֍Ⰹy0Sߚtf=VǾLu/ɡ="gx8oGus}1oz	uںdt?fn+݃s{{=+e?4DA:K_۔o4_Ig
'o=+Cӛt=Ml1WpI(/w)\?K"ۙ	-zYdJhyE /{ӱ%4=^v1.siȋ0(oݤM~ub}13&!rFe}O/pp'Qi4L47{7ԝ4Soa[ݭGiqNxye*ǙUg&g3|51u>Y~Žu{7YW_wذu-.b[h4Z@r$/jLj0I6S{E_duvb!兔('Nhmx !jI4P 02\Rq@QDۗ,zYCf0JbS'USoϩqQʹ5nVRipiԚ:D58p{cոyd4EKb$gXl-p a-(K2@hqgiwbyyw[l"!>ia˴lpYu&~	zJ4D%ly0J.9 Q;M܀ iW^(\kT*(w=q4|0%OAhntIBD2VȀ1(9E|PwNBcP0!9k֗䩍uDHʨA4;Nk-ˌcFYL:#SO,B4=ҴW0\4"TԞ$|B:(/sB c-m3tPN($+!8)EmR!z+t,o#2t$&޵q$2KlFEp(uE}81Z(&)"T̐%R9(kؖfzoU_UwUէ
4cH[4;KL?٤(<~G&G+M	/&wQjo'E(j;-5e{A{W.09Ӥ0@2 S?r8B怒Swh&piׇ)0e_	>*:
,%0tHp7Bj2#_øYGޟفcn6M,}֩p]Lf"D淺̅G%#czMqaF^9]K|&ȉ_r|6QǗwō>եx2Dsu~~p Lg\ϽAޟWcK|l#`8Oa'U-q%nrU3de3ra/7Xބ4f#X9?]k:'Z*A[m:U}U٬I2fd2X>]ja,8‡VV_])_W5񶪬7^7:W?~}~wޜaN?={;8_Lp_HZFxaɏwCoҭ~Y6E'K0/|IQg&]]ʁ;ϛT$V;M+V^4UlY|ve]^_U\Y(.@@W|\KZ&S:~XuMGS艖VG;í83C-T`޻㝎Ժa.e=ym%:p`(3jAG1S6&ӚԤY%BݙN+g:{'ښOV;϶';{y0;P^yqg4vRijYX%W6uJOcx4<0Km*CL3ŃsDXMƭmNxvYcZ+ΑQVK@ F8FPt';IbQgT
VqRր0vЈW逰Dd~rVkLSJx K@)*2Nj7i[QDRI#ߛ_4篿Bڅ`ֲ[ϼR8vБFӴ|&ø"f?oj9-J-AƒEdLwhHdžD H{nј %Rk(IM0XKDkQ1I&x9Ȱ,
Ƹ#aR
P)$Ps?$ZB_@]Z:އZ|~PB>O}S}ÐL#7ԑZ*d_:DWڠ1y
y0[k1
EQZk}tPR\{ɩ3-gD-2@tGe[Az`q*[cdc?yw|9Y
	˺|"geF0Ng u%NsT@"\1y?;=YaP6HOC6/$e
IyY%Rh~>XD'&8f|^"$wsy|iURn%H葎{yg<)4aR^X8ֈX;ịUh4BkM̨sZ
ۨQ1,%!	j`SEL;5rw/`i\0{lW=>wN55zEY>f))ku>4gLꥄO@n87z^郃Ak.99HI_ɛC||'!lTWO#a7jzXj;R֐0})e9Q0{),Ho3SOJe;֫A/`2owwкM]nuٰ|Dyw3Rbf$븨^OUM*ޠim1ʯx014Pc\Kr2Pv9p谀X\3RW5?zKYf%4;ʔS@0X)%;ZrE-T(lg
{zMa]PGC^0vR)m1HLeHIj숵Y4\8em>^g
?Sض9]9l1'\'s?ݤTږ8(X-l4U<U w-–C[pVGM]Dey "1r|2㿚GyZGyT҄64\ii}b%fX$Rs}qy&BQp񵯹~6xUq}v&Lߎvs-mvZ#jm)sy.>	A5 SlZ"ۃjZE:ĂpLx6*KsaWZ]8:vؕќO}3R'-FfLqh]Tz7dzhvUNESO1Ōs6[_8{.
\z"ckOg=QKɾ&*_\J~ndiy gȤ/Tθ8sA::KL%&ĄB M)t!0@	Y1( VI`Msۍ.C0!B*Caj3֦4k1hR!--_FΆ+%$%Lףaof܁bpڊH7oPL8fLeu&uT]KͫNCZۘȵM4QWLWjPC TҕM%.t*]VW9S~Ϯ\!,dYͫ]7jZY٠慒C-Cy7Y_7g]=8aוt)z:J맮Y1Oo|wz!+3A.hn)L[͗.o6iRA!aÇl1
Ľ%w>n<=~<7#,F%-{5)c%WBrŊL0OLoq3 <6l9JL2C6D:i<+hA*zlPO"E(=>uو/^*{|0̈́:o{t3${\;oV
'Eٟ5[</]{\p͌?tᕲ|_ٓ.ٓM.'HӳEbX.\`Njqu_Od\g:ϗrCzSF4a>qtycRR^W <{';4O~ǿ]͗}/,cmJ/FtoiJg~N'oFx򗿜i
;	:<TA!6>;i-C
rd[c|~]v.[]l
V"N8רͦ7gYx}v<[_'{ǚU8XƍgQ$1LOߏ.%l|bO9ۨ[vU]t-	T^_zO3xBkXUV>y>Y4[83?CDZ85P//?In"V{z#(;4W<ī^]۾Q-xgoKsʍ{Ib
1ķu=u5b]vvÕ$[V,bwu1nB%:CZUN\gnA;Wܡ^yWWsli&Uk8/diPbdXwMhu3'SӇ'+(*1,0uN|2KJ_FDtamgH"5ڐkTS.F KK!^c:a`T$〓!D=,hjw3_]~#mO^Doe7΋iTw2>\.RwہXB7RM˾4`)[i5k^Lü:\/a^m}oǣ4R
~yAYcݡ歧uت}iP4A]9vZT78LU{7?vC;SE	"	e0(4Y*X5:;L!G-~$ys:-g>baH+.FVHC9敏 F2u^8,hXtN
{	g͟}3RKAxtd(Ixt<	HhD2Na;m4i=.z^!_o\󭦶'|=s3z>_Pv]:dBKVj6h>:Bl"p!y8wǫO[v7Oyp^?Z!p᷃9qڪp0}1il_Yɘ
9w
-d#iJg])tEYQ	B蔲հ^nϔ%3Drl-8C I}DE7AHgug).Nߵrg;[˯*u2zq/[Vm0THTσyh.Z"ni.|_rj!wTbجB/-] #uְN.J.&G6Z?X/g/FX5$AQOjK(5&LRI[J&+adA DO^"L!	sL)HcH"DSƒb1^76Ύa5BvWۃHkXҍ$*U.lKc)Pt1FdGmZv2Hұg+O e&IqC{oK𹱣6J52-i-/?%ŭfZ,lE~]l9WZ֦evdΩG.2dax"Y&[dȵP`hPf-q
bIIoW-h>{XBR	})&NG)A526؎4f/Xh,w}}L"MȌ F
"#UlR=QFjV*m:DBQMPxHWLqج+4W@kj8(ԧT='~?9@!'i]Ot$8P<QJIݜt@@&cILlS@d}g\]DA-
T@YY)ˣY5c=4<@4"$i*/6<k'k*9X1[Y@@FB՜yh8j[_>a~rrC^2.((&ޘIR1ӆq(HB26SX4Z'2/%R6;٩J*
qg}~:İ[?O%љڽ,~YNmf'iA"O,O[}߂vOv_Z3|
rOҴޱ^Ew'Hw/e\aj5>=[~\?|x7kXCs~??Y-˾][}lqsE?RC-Y=[:j6j3?%|86n8`FE:t/Z=v>Ga"_͎LhcWr?GKM_?.lh쟼A8to}!}^|Mx_2gCe/\2W~ZˇxhY3X}Ʉ2꬞
ܾ:+ϻw54߽j7M4yիK~Kh/W
_RI9
	܀8a}!JI;Pw/#yM>VH>lGj%c^M9TK9nJm9\9Q*E#}ņsr/f6qymf`袙rkL`,F)&72t:obW.NtFsޟlJbA5ZtCl<7r(5jzlqۉW6܆̭鬏dsœYٹ؜gǖm~9U䘾vcjESa9'(Ew
[MT-YQI۠۸gTUlJE,͖YX.=
CgvI
;}&1>_yrվ[
|HoM:zuI[T2-w^NP8U{THykJgpQm:]%+5+"xmRW8ĭ+-M+A~HWԚ^sdݫP?f^<0:SvҞ~gϖM;ziig7_~eCDu[ܨX,
O0m0om%;4]=IrK]ۣ7~"#ѕctr"MtEtAoKa[JІ+A%HWDWl\F]l6]=ևMkW2Z57D6~(DW2wzJ=:]}l\~h#|?a܃DWz~J;o]	h7ҩ"]xԕak
pm+A~JPz?+R")헎ϧ+ADWazqe1oYp\(Ϭq\1.lF(:sB*N:eb涏\-z:XvJ^:y{gxGIak.VRoU{J:Ie@-2BX}
gJCs>E=$wL{WvQ@g˫#u]	a\U,B9aA}* K7w2073p@,+ܳoGfe0ρ>`fE!u\ŰCؒCIz.PMVWˑt&݅]щoUA?XG3x!]E˒>kMr9[,ٳTۯp99ӓ.JE]B٫%S} eB6źZjYF5PBRU|)jJ%#'{U]*mg9b5Z]ُwhQe1RkPTA*Fr-6}ԫN=H)|6n-6P
r|xj(zljm-e()wfgZACg@וl*I0{NRi_
>M(=:fDyKƜ9s,~JurP/쓜|weVU˝=*xN5thN1HGQo?moaL>,QێMѵ5b'ԈShі8:AOkc0)a@zQimG;&1x_I;MBP`U}
٨\S>EC:V)3HT*Ht=) ;*D{jڐ]jy;J2](_ "

]d0`g`jb
|SZ\KQVaD9qnB!VˮJW<.!kBhcnuEѕ<q!YGeQ\Zlhu7vM\s	\/S
yt_6	5DnH#H܎m }K7*YȩՏoT?/Xżm0|Y+ QHk
;.}2di*Yrѡ",'Kе	2"1wPڧ5hxC$*Kt]/sN3uk ) b*L
`Q%hB;SIJv]CZx
_3XُBjF,F,u=`8,,@GJE1QIb&֪(Y(cV$C
VAõj
`fnqYu3%Vb|؀?vX-+,I52
j3+ITʵrVnU{9Ex`!-##jw$aU50l#/^Ғ%Cj%M.r+y@MrVkA\cUw&*>Y[8mP zxq	Vr&*M̀|2/O34%0LJKFD=45QpJ
]\1l@b~xp4Ttw7j̦r!pUX.
Q b!C1+%Ikr)IP.X,;)jI-,
v6q	X"kpN)aAAzK_>&T9Y9$chܲj%D sl\}L)s*]en`8^x6[Z|_4΍84д{쪀5qv16nN 
fkv/r@@I'Zhi'дh	4vM;@Ni'дh	4vM;@Ni'дh	4vM;@Ni'дh	4vM;@g6B
6cn&vmGk44'WEyJ#"c!ͯu@"%)H	DJ R@"%)H	DJUfl@9֟-3Q46)q3xu__hR7
\W=7ܶh0[0}9şSd&q((4Y]C/3Wu"%x}r8lv;Sz??˞QkVBRA,M	an]bW(ew!#Y{w]pzϽu@'1ׅs^z^':INuRz^':INuRz^':INuRz^':INuRz^':INuRz^':I~<;=V)N^0[ݤyBsyE_f;@-W5p	2=$W8	At!,cO:Kl\o
:H]`CWfSsD;]̎5DWGHWVaRPB(#ⴓ]Z;]a^GIWNA)Z5/th?w(]":{DWXXܾt(DW_
]
^2>w
`#:0]mW	sX
P߱v]Iv=7LI#A
]q#2]+@>Atu)^b2EP(dDb4oGlgc"w%TE<٥1Z/0x
_*[/Ҿ|ޙ'})K
njX.+?w]^xM̳	4LQ쟿oт@zn+eW)7
LoFjPbnpu͗_P|9,#=~	ne}J)eUMCK:|SJٝzqE=y{!JYvOVXUe-(dV2:'^wd-u.	.VRRxJŒ,+C)EnCKw*.he>ÿYr:2sU*2_V0q#$%*^7[z3puu}jQJZ8Ʃ-}+
]!\gBWm5ڪDWGDW۵&`)8@pF>c9^Ɠōo<(o>Y8:}}RhuN::&!Cyo'%-/;=z>&q_imw;/?swrrvr'vINna2ԗҜ
šƊNQc-{2ЦM.m)|{q C]yr9#؜?%' qj5y_\֧Iq_Oioͦni}k<{u,tm}:n
}@;yؙ7 +Dҕ3k#7CݟBpY
"\'BWS툒̆]oNZߙnJ@|33_V>o0VhIvq{5Xg99]%!;IܴC	ގ| n}q
<CrA׿C;+Umbw/EJ&7g8Hoʫ0_f7
 5sƒ1\Y麬q%Sr|?ۚW5p|		ْ8BhJ!,o85h/#x?0CE
:ãsKhi
B	
RXs|t=~8)-p;d/<>|w!n?$qih홟-jNS"%,9%#"f*9H^-TΕ
K
,:`̊&a˔ߟU^,"m}OܹtpϼJ\GԀK|TP:U9T0^`Z{`8h/WKfs^^^1Yta~hڎXpU3la;tu<ՁE=i`ս-fzfWgM^Tk{jB'(;|/_.53xǦg!k=W=%'t޴쭿w:,^Oo>Z(#O2܋Sn)R}0zyW`+F[C{vqSh6v[?^,wx,? Ba\~b}C>b<~W49]T4D4w,8/~.f)([%Ϡb|={+
뵽ϓA}tx^|LomB!\-/
g{xIBBxD/wc!w&|ڴBJ~\s[ޗ4|aMCt=pl]XabT)*S&ji$"?~>'O+e'9yCNߓw>{8yߙ
^GXŸʊX赮|dJ7s-r)("ҊD	K+yL)ggŜ2Z$s|~ikxԒeE+F&|B@ЇrlY̫sGBLv\!&:<,;<,;3b*0dUƧ:;`}`3^W5Eo9lt-s>T).CUyT\։
^p󌭄slrsTbt@hPVV7)W7b`nD31.zZ?=Aɓl7Y,	#M?_OT=o^j	OG:?1GMQפWǫZ߹)c>ރrJPi?.ҏWs4}eՂAl*%,ֺy=;5
?]|ZxJ>ngϔZyVsOTBvt?qצ~$)li/n|vP̗άJ݇>V<'7g#7]r9><~nDĸ{ۡ	M$%	^$R%%J.;¦X,TA /۽킬;m.ǿ};f$8f\0&JÙ	F
ۉ?9'OO+̞9Lʬr#s`H6H9U,P[ *KE}ʦ!B[0:Y6r9(LO0=%LLWM/7T˦<㘠vKjx9FW|\N~Dkm%,uddBE2BtV(]0ZblWx:[WOyj`P=Ům=߶2qߐEL}\2z{˯y,\tT*'}s
Ӂ7UnTb:&-bR"Zt-yU_{y($u1Teɦs&(Kp3nJ8G8:q^5x-]GÅ@z9a\owm$9D;DZs2fǕƤ@zz/:0@-O9ԺbM, :;I5/@$*X
&j2*j5q6T<-y͘RO'kEܚE"?fz>z#=pT~xZ#BH61EOҎ!rl!1IY}TmLh\,ų̈́-.O;jG<ܱ b
2k++_P"/:9)׋Rz"mƳzrRE*u6LTUVXɥֺ>7.gp<=˱>cw*>:Uoi?׽I~Ǥ+cdDev92a8&"{ʡ"1Pz0&Qw;%.)Qbf]YR	F%08329eL8DlR:;C-ߋW>OLYk`W7kf?j8zHG9E]/>,ವBVjx1I/A^9QC2xØ&N\2MRAH.(E-?wKnI7Ť
G=Y鍺#it'ݮwzF;JLә	t&"Li/w~/4'S$t7Z#:Nv=c
?O2
f~1
Ǥ~Da13v]̂;ivE:ot~ia-4mT0q1x,`w;pÏ% f6kw?_FY.<4e;5ҴD)QgU"~t;㨻7i&'	0H'4s2s=ڻ(m	ޏ@Y=0L]
錌gQcI07|O*Ppxy铝/t-IsebZwhTqSf~R0:u?iar}^-7Ҿ_	&dy~}xMVöpѿ܋AYxwK>&BeSKm4Ϳ-tsp6,T29k-,vɦx&4Oarp8IcUқg쯰<;ߗAŠ7Y̄ȳ= /]?r=?RF*b,"va:EnxW,sd&^ͅ-;//hݩPfҮ	&(kZ?\7j7]3,-&%Y*[p\IᢲV9n`e-:)`]۩D?9X$b*mU*˒Fp6X:	˄BBe*F^jv+i3PQTӍ^_xXuȵ##nRCzR[2yJFaj|7q8h|~BvV2*i,<b Sh1eiM,lau>W,SoRF؋L~|me7{vءS-˵]M%\c.FL6c@&xaE:qyICLɐ#V2>5n$y|-b٦}mKah獭OaZh^V4(DFek֬('U0o'R=^XJran,-hLp<3d:ru<+2'[M-<7x'k,ݿoiN&\Li[qO@mi䜵IL-H
]V3ȝ)n	,O:G797Pb<	%9qk%#UNvjbk%QSo-|&fYS(^7~ 5ʃ%WI8WƀFMI6ɨ!hasKnIܒ+OoIGcFB*%䕃	$bTA䈔rW-۝[>dkɇ',W<) j"F$!LED
I9c"䗤d-Y뚧=7H0l]3Ww⼟Voyuy{oQͬNFqGdxϼ!H<
-Mxn[M-+z+;ʕ"IXh'ËqM!w~|6X6MeK>ײvpA% iȵA	M!y13b`DzRC4Wt";7>Ug"aIAANIY'J6[+h"$ѡXUaVx=g
E{P<IJsV\ M.:S[d(DKӌWܪk)NMO35mU"IiTT0V*HȽ16D䨡M&R䱠>qR]C%A1(0h$1zI@BƲxwVzn9,͔
r]B?u;UM:KBu11
%?ƥ	tKarnuMcgs!cG^jf5ċ^bP=3)KH"eA	θdhĎNt&_&SGg)Ҫvʳ	ew֞iqvg~zYZgKTف?_L_]uF.;=c8󺡃ӷ~Wئ;ͻy'MocWmV?{WF
yXGDEgFc09HH[7ET,ˬdV_~QBe'5>Q^,d<;نin־Uo]4r[׵o[e{?-tgJ篓P4MIx_?nV*U4'~o*ǃ0<@/G~x?~?~z/hQx]Kj+cXXLwpʱEԍz+_*ò=A\|Z@dNa^ڎ{w߷[znmW֫}WKۼl÷ynk5/$B@\^ꃗL$<5>&e#ú*gn#%)|joCr!&oY2"٣$wFm7Rk"	ɂҧȓ:ZA0>5+9xnJN+g:ڜ)z8fe};EEd<嵮<]\ynUXn2<ҙB5}u)Lc:A`oWz‹WN)ɏnL]1N'_؊,ޏ<$
WfOZvj?)J4#%SLd-jg"xE֛~./N/.1A>(KP'W#!=MCE/,GN^v~8=[%_M
l3v~n>IT~2x@-٧z?
xX[0IUEoIgȥSBX:HaA3hxjլK;'579^QG5kU-6dġ{?uq.62_^3];fgWk۷-@}z9p853(1
RoЀ9ϙU=|BzElWପ*;]AZdwX~&nͮ
f|* @[쪯o?e2earڨ,@XYPcd8g~䭲/s{
g:.Gs"*qf+ޒXMsi.(Қ )3Mlń1ti \i:\(1>GҌc"UGW$XH{`*R#\f
;&cJ~x[ IiY%9•咩c+sv4pUXH+mH#+W+#f'.DžI]OJ˻Wp{u`C}"UıUUR!\	!8&vU~
mk?*ڭe*R3+)4JyDpEzul/"iu*RpJkܥXqQq=	Wd2%-_uvP̚փԷ/sU
_jty/BG`ܫŧTWU41K%Ir(Ԧ
cÆE
[
c>$nv̑~1	E60ǙgAWpAE
Aj8L\+zb>EfٹlE.cy-${H)w֫&t@-\HZL!SV'WNeCD
-":!;OM:2k8ޖ
oa[To6w@Ɂx|bp,6EZ]?)RZݟ;=yK^ס'\aLHh!DtJ@{}jz}%/
.OzO6s)dHhh~+
~(vKJ;'2wdRAo[Z2dYDk0D3XŕK}ے6>}H+q(ր`4)x0re] fߦNq{}}_7>8\Ћ"os@:n!r3Z,Jj?TċQ/irBfH6YJ{JHJXpjy6z4xԭiUgY=˺emv=FŸŶgZĴ,3y3lN1cWt'WBBH:өC]VЭ~ْ~hJV]Rhb&d>XecNFu[Ƴ7qHKIaoM)z༷ qI`39rSڋZgK;'ο{qk#ξn(糵[dk|
7:Ӫ{B]geUOfukktlk!76v\!B.~4W$
p[O۩tEUSzsԝ.ˬoͺB*YQj>[vSިݕϯlQJ|2Pf6o3'㼝=wT5)HѡV0ep'mHs&Ne4`W(5dϫ?3	x-(=Z]]/W:<6,#g8ӈڡ\,rU\ԭ\l"Ey^'!,'$zD'$	QA$\܇\<Qz|4"np"JJ0	u~ gm4}}~EhZ~?
8=
SS[k]8J,YԹD
Iuz2w2~rނ.j>7F24)Ҝ 6R+H(&ʒaN@RU/	J9˚/co^t*bTmBn`*QfB@'DLZCcl9F,^ÏզtH),v.XG6gzM6N1~cbG?$`br
CZq\*&v$9b1K8~:%h%^$B^e
V0<`%D%O.heQ> 
Q:TsavtTSࠝnV0|	.C$~x}
JO-2PO%Ȳ[ܞ{=E9MIx	w*(S):-r^1if˩dR(PYZӌ=kEO+y>7ɓeDFN8bz!TBm XꍈFEc_JF$O?	'"LkŁ
4Hq{i10LY/C5<-h8d0tQzv'!ԄMm7D(%G‘8w!7/l^<נ_^>Fy_߲c۱#lӧiBRXKE,D$:8sϲC̚?$[s_S}D'1X5b(gBf_#5`4{	']~uF+:YoE.h8>9~ΫZ7S:+nk֟"s=il^<󃺯rNYqD,3Mj:(ΏNBxSQyrᨈU^́orxˈdqzӗ/ԄU?ȺM7
W9r\Mvdv=FQِi!YTP2)=wpn]O]yW^m\#{jl'?R: VJ_b<٠7왦<
3AygJ-K/nZGXqHy"amUux`ðM؎yO{ä^&ZYi-+/kYΊ	!фͤ	jZ-UCE`ǹd%:!5(I0|Gg	)XFNeaPLz
m|Ej@/@v(#Ӕi̓7GbJB!(
F
Er@QE5h7#
餜lt(ݍǞ۷Y\xꐃ`+RʾnOjc([+Ծ*,""x6?N@v	Eq. z[rf]%O6S{+T\	{ُ~^&w`S.rhT3yiGKfwM`s\(K!E	0xH9&=E5knu@ *8M%#_oB?dh[aTE^)4RsψrԷX0pfwM$
E	Pz^JpIԘ)N1TOD=M*"DfO)'K(&iPsXd)gSXeUxiO.oV:	ǠPAUD*#TYШI}bW#JaMKYj4X9Cbf3A>C5>jcr磶zK>fu1Ǘ-`H5#.w2*e1Oiq#Zۑjq8ݐ;n9|\Z.AR!b$y@ཌྷk+ b:`||D9ϥFƈR.'Np\	*D&
Pã)9[*Kiz0Pt)Wv+W,jÀ3?>9
[ܾ..];ަFz]Z.=pY>r=0pdJ1BPh.u̎F#;3??LQž)kOJn+9|<o&WaKr9[w+छWPTe(UosPXu]Y̻p?_LYoap@^bz]7VQ,{d~\HtϷñϏMb1LEpMZ?7o7:/GoY_>U2}p4z֍K%	KYX`A7|UB;aA(+8Jg?ډ1w_׺wIS]ct-|·~گp!JYDfoE:xIC0]զ7G'HtQבVS\
̀3DJgSb>Dr'(5Ց76,KYq_h^;X5Ψh4	#xR6mTXPL"ʴg:Tckbv]d]yk]y:Zq<3mmgRz^(vվn|)45(D*պ2lN	)ksorSi;ѝx<*;<"~(bbBRnCA<0M#[QhJ=v!&ji­ß3&(^U@) JNqRcl#HyCМT(7ЉUstd۵{>\ׂkn5wآBG7:RH_-pqo*2PO5U|YM>yW^ɫY©p|~7AC\~ιgx7r5w_j[Fjy"_T)‘KGaXY]"g]'[=*^틈[M:ozo[+x$/?oDYB%/Kշօ?ww~NzE[l<·_~_yh++U#2Q8KRG^>e}^( |\o4ͻݤ}@6
oTɴ楰ĕ:V}kp4	@2
V(c 1mo@
X-rwLṱ1$$AUI	!=#:j֨d:Aȹ]*>:kq$_dJ+Rn[M45iC0W/${YYpoeP@<)ʛK
e_DEbui'=&G(1C<
&j+qT4H;ilBvR0R
HKBA(K`JdT?)_?m4Qjmve]
\-| 	vYdSJ%HIJj&R𳅟-S,"Sdu淿\_"D4B9˚GmcT봚B'
!CѥwxP=UiFB`"HyeEiKQ3Q"RPĘ1h%e2
D;u@}
gr.9
/e'DZJ82:Gj-Z;v@w&'\irٓFy>8|~nB;/<뼹]_x̻;'fWZFmG" X)ͱ&wc3ȏE4v"Jocx,!6=rqʵu
jqFфĕg/w&P7jrm7NV=5<WMt+•gӻc6	W:gjZ%a*YWn܄6=iC
np*WM㪩4v=֪pN]5zUSԾqTnpu0RF96	\A'Zǎ儫CĕutY}Vi}C+Siz2~cb:Ѳ/V'B~+?vyiFmq}	|Vw~pY(r?^ߞ--}v.r|ޓpXna ],_eL([Oë˛t~2w
*M2=0nl=urivL7VM1ӚȎ?+\5\5^WP鄞pu2ƛpeHj
x;ґpu
u?m>%FX޻JI!I%u&np$\5ƎLK+/=
N
WMƻZUSvׂ+kzpX}j`M{>+9PǸgjʑdq\-p'!>H~@̽઩jj*G$	Wσ+q
վ`p}%҉	W+ŒMO k=B1v\5#K.	WυvH&HS)4!r",WWYJ<WPʹt2
OvK\5ʏWMVW^;~a_E3XaR7%>K_.^WRGwm7?\7o~uSjL?o/>7q{[/}UyY:s-|ﷳWv#aZ^v2+vQA;'V
8P򣊴׋ 6g8K+:Ҏֲ#Si
|t:ՀNі}.[>-r}~zیlKnQ[!-o/(w|og/h[nZmwڮڪ}~d@9ٶ/XvA~Anf¬Bڱjm6^fq09&WnpOMtJ''\> 8^/$+m̖|;~ZS^9p~Z]/_\Ed^sQW5ZZr(}Uv
z.Y4mnC<+
^?@8o?-/W?pl~Ʋa?E[_$t)oqp-wS+V_;7T|ԥ_κ~[u+rlYYxR/{+`|bT]b^_߼Y~Sy~@l4aV>n>͟=W9R~1>?!NQMF^{Zuxۦyͧ\CA)\~2]c16'їxBƩ4eDFX(*Cj\ (([KڝnNQ~v%I=V.q_Sm—?/g
/(	ᩚ칸;")DMA`!`;mFs=.Ge%PR"dZjb2e%…ZMH	ڎܶ&-
I|fd*rf[0}{5l[$˨oUD4PU-{GB:6TkK	B.:d91<1(FBu46JU-)5䶔.6[U딣TV-,.LZExRX"D}`!{җ$dљьUJCfU'Q7*&Yk8qjێk!c5GHE=ᵾEkѫ
Vgkn)$RRy8Uz+#$|0@hYH"g]H$V(iM,`#<>m\)
Z'kPV	zU"+Jld)j&pN%)PFk@zA
JGHRkI풇ɓMD@(x1Ku-%z
hDئzKaH	PK[#.&%ibpE"CjbR))"(Y*L6xO$sCP fX0!o+ຠvZ$J!BKT1n{0I,ˈwʫmbU]e18xBf	/xtT[_+Kx·3l8셲;5Oǒ5<۞MnU9@eQ(UL}
MĢ#.\֝Ҡy\ZFo(jٲsL)*.A0`ډbvka5F`ex2KN\1zP2<(dJPLiX|/ݥ؊zTuVBXhb	Xz'ciI
8\w)#DPP_*Ӕ)ExF2\c馳ȀdXRSU.402cl
q+)fa
¼uD`HHH <,^ TY')EΥ*#[GtXYp촃w?K91۔|I0iv
 b5V12)_#LVJCP{sv	a*3j!Qb.1(c*Q
LNk=B6:j?@N9" REte"bkۊ.ީ$h}*)ֈLVXE(mu5!`Ux@V`^b-x">R3&hvGq3V/Anu+ZKo3f='EJvNDu}kmȲ0vڪC@Xdff1&2 gL&5"-Y&ERHQR+"#.ުu뜪{fغ1A&[Tז5p1nk.FU015ZǁmZI|4x	 #.*҇АJ'.-*=``j>pLq`w-ȗXP0F4&@21eW$&EJô6VYѰq	
Edup@Zi29W.nc1ΠA`&Ybm4 sp0(A/1o* 91%D,s
m堃\
Ec˄
A(P`AH:P#fi'O
H@X
7Ȱn$>G/+lż"0r\O.DNa|;)E^g<[0S!
!-JFJQ5zg
b"Zq܏!s6,ѳdVl,2aFv#
|,$˃>0t$ƀ?!\,ݲFrîAp^%SCxYk}(Ѫ`»e:rn'I#,+A-t1@+hB?ܘ#"8=ZMfPO,[JMѓJ 
eK*ncvNA}6DU˥71.ń120+&k	7Dv.P,9)`$҈56X~vW$<Neg,"B#*ի%B3M.nT>7,
&.F;@E/B+yLAb1rAdYUyM{=R|~l(bt	)	vYJ/Aί)L&-Cofe^_/8bzV&ѻs]|,3.k?42/lJZL]q}WG}fp.BVG},UQǾŨa-=*Z]sh=a΃U	)ޞ
v2oi>mjVJCJBR@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)\%`[B8dNf'_7۬@/Q	)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@/W	씔@+NG	(Z;z%n@/Q	d6%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@/W	Ƞ:%%kkkr}2J ǀ^	Ӝ@/Q	@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ/G	yjew*U?^0m}z}N_%~N2~f溜G@!;%tK{v:0Bp	VJK¥/F$zEW3غgW3׹
WV~?ܮpu?+8p%$v=gdž+άѧޜJUR
W/0\	+cN(\>LpKj>lg^`MO(\588Tj溓AWɳıfT^brR팟W9
6?6+w-A>{>zk'Yp<_t3}9ƿ17<-fyw=v@㛧\oբ?0
<0UTt9tZ
/0Lk'n9'FJjֺW3B/1\>CUy{:kv\~2$Y+屇f^`jGR)yB
-~᪙ةfNSzI'_Zv.Gr{_|*\óYo%Ϯݦ.?^[BXM];ٵsg5]?2WWX}fo?FbWvU^{ rj]_}e7_"TVYYRqJv
Ȣ,2MHM/(iuڄZM7OFϯDpPyVGoӯW髜].ClKw$,?lyNVw*'M!UF2!ܜm;:lg/s/:)zHw?97zfd[C[$9EcߖMw=U3w	ɪ~qږF)]Bi:㪴bN^7+E2!Mt\_TQRns`k;acVmٜ_V/XFtꌓd>nЗ>xo>/
 _]RVo;|}gXqWj/j6ywm_{>0j]ΧY2NiF'́Y?o]_n{}E۬Q|.U
G^jj;ZW
Wng*jrE8-2r-OMoK:V8u29wFフjcR6gmPp,Ç2zQ+361R昨踬|/|ƵUNb.r5ۙLۢ2U05Ĵ(Bjf­2g51vDc첤Uy隕@/rz^Wˤg.k\q]tutU"Q
AϮŃly>e"Q7BY;Y
"HĎK_RUF+íĮ)YRsE:MʴWuؒxQvWoQViQ:@86~29 zz1d43<0J FJރ3@\#/bՖi<`9%QhIQKcvYb`X9I^\[̲oa'N&VRĬm>.bm#tQ.ߍ?AʯKo_\W$\7|6g.&eU?7Opoyi0uqe&g[r-7SuˋI`gyNI;os"w/˼,:)u{/wcn_<8V\ܸZMśi?|K_*|1;_t~
C9*rxW?w/
xu jم*nL\ۦ駽2lIz,ɬ0YFRB)FWᘋRrT\ƜB+U0e<yF1t`>G'	Lq3l0_޼~@D[kޙЗj,|xѳE_&Zmgl^IEh@hѭԷnfTƻ~ww[޽x4t~G{v5[/wv[ykJ|w+	%oݹ.g=]_ݟ!>?Ҁ6:wCy7vK0ZIMZuo=4bqOw3,]Xq/R;i*:4܈?7WT 7:O`j%cuE)Bg8*^gb"x8}t͆qUvxJVV
52
,K
*Uy.~UUi@JRD,rV%')jP<$10=̜)beg"H5Z/jDTG'&=w4npҔdj6;(udd**0[25pJgF)y5>9z;+?Yզ'Rg-wP-
;7o[[7o1"If6돺Q/ekQNj]wUhGDܱN>k*BW(M]Iٻ6WxIl
b͇ hS$öo)E
=6 Q3隞!:2DEp6h)Ox4$`t[W=_no
lm#|=,Lf㩵h&?ZYCovZY$WSkˌ>0+Źr~E͘ϤŸu)c,:\(q%*r.,+MeidY֠-mY6*eDFLp,h( ƨKJH%x4dٺɲ[ Akm@BoxX[94zAג1<&C "'#$4@Z
іCeA$O}C)"ƈ
AP,,m$U[%mRޱz痱5Ƃ X&
H%t-L%ʬbN P
89jMIcxTbOv>=se	hѻ1z<8{aRnr#~5_{cG?#H4@]BSXˆ3+H'4)I8b*pdE B8"TqCB( ̲D13Oj7[jo!~x{	JOГg;TmKvrEdH4@5h<#fo1Mb8dg6.]Ț:@<Oo']}UB,aNb3o"'`i:re -"}w}i`!`19"D(' l[JeG,$RNMW?dpT:}7э:n[Um+\o<
|noԡ1$td18\`8UBV&PLFM}w.+6a?
ߦa'CzЏi<7#F8w,!78נl揉sr7>#Aůb;EѝzxKz#ֱR8eJJA"mEe0;8[Nd[N;ϗJuGr&d)&t1ͬe$Xy]LQg7z]I]
	\9>yc^xy@]o;@Z~@K\8CYcM-Q2/oF'?[jelR7ūp48QMcxS>䒨U~(חfp|s7k\F@g.Wne΃_!Ykg3Zmo2.dAOɿ14;hl#x&4v+Z=rޱ
zɁeʏCL9;;qX{֠W׼tq:\|%/.T$n@[@5]`ʰ̄]m$AƤZZ6LƁ؉dbJQr
(.hn
!]R(TSd؛Xu$yhcݩܡ!2xxrACMmNEsA0c$jԪ"h34d8h*)1@-w* D-'1Q9!c-m5z
iC裐@JH; qJ}eрKC@H	3)wgL<5]x;",Yyvշ&[.&',ǹNWřD|X|R#%#$q>xS`ޕ0֑<3a402Ѽ0;S8<}h2E:^?T
f) tv
(hIM|ccYްzܽt1ϴ_jՉgɺcv0RQ}
Rwz6[n/vmr)TUaT%I$-okmmFf֕b!aO؃(>/=lwޛp2rsˇg57U<K
|;|a*2l#Z~*ԉ_7
構:䟯uo^u)3ݓt_x#?pP8\RD^Xaad=
$eenT5g/C^	
b_'?B<1rZCOݛ6e܈]u>wiW[jFK(	籂Kj%Tj4}c|M)Fjg_yF94:N)2M..%"xc}8A?y2 ymTd\zlZsW$F0@)ؤu.SZ]V4(ۮԲӨtb]weTՂV*ϕTkܪ0J\vU3sv@v5~x387P-vK_xR|dɗ<}pW]C?p|q/smCrﻟr%V
'5.'T3]T)QJ! VvV
8[Ox.O:sГ.(1xʒy}\x:]\I~|8w[s5W쭰aYow\Pwg~N{e<[l<
'//¢T|&!a^\
V5.
}/ϣ.={WorYY1dUR_!6RNS
C!Ri1DQz%04/qWlBCrE:՟ {lA[	?ǍYa^N5gx+^Y'X`o@Yj+$,OuѥGjHXVS"YUXUd)έƠqLRf]b9ac_6QXFBK6*jAlTDFy[6Z#yJB!R
V}[hgzO@3QӇZ&FKl;cy9F9BtԿ7yA^L
N7F'yNhD$:SA"xjHO6Zd,ȴ57.jgEIn	K2E֥H3/fTYI	<DmkõT3j[#g7$;7iEOV#{τoy&ZW֡ON)rV!_IJHM9,_$ESuSi/G['VۚhiJNR
ŝ iRAIHdpIJ4tŠ7*?|
O`	f$ϸOG˟
-u~r
FO1Ry$)G^;7?
}LG&͐
Il@eGK4/{5NMVxbtXi)&-tPNqWCؾ]1D\vNwŤ$3DjoI}dHΙJC34Hd&Ā2	7y[5p毺jGJU—SiQ!qZ̖#zMxU?.Qd~nJi᫢<5\U͊D`+@ޟ(7XDj
ssv"؞WgjwFXSeX~Ưz׽/g+3~
1_moנW65}z48JZE!Ҥ*?o
"
A53{7wC\F
jc
d_TM~ג0LP6P4fXr&wRg!MT3Cb9H8}U2d5\ozIHF0V/T&qVSghYXE+cgPȂYhZS;ynoO^kU6y'`%x'wPة{'{wx'F#!g?
vLݯU+5=ĉ4A`w,Ȇw܃r:J*)ed:lέ2!0jCֶ]m07_;Զj,JxX3)WܻR
XvH$c9ZjE{.XC9E<@.uJgqєWY~8/[8P!e]=GK|D/(?Oky{qj~Qwy\P8;2n6{OW+텐Z2jΩDA.%DςQ5"Aeʽr-P}O
՚K(/f9	3ͻݵz4g
sY>͟^^SQ\0]WFiI=HQ5)ANP
R`SAZ+<;e7ҥEmEQRmᇀ*@?_ܭl=+CU>zonß]xKFǏdsIo^{|QRtgB$P4W?_#F"?!ثy*]x}W+^w]*^o)GWdՆ̣y(:"/"5"{EY(A9O[`@AҧOy%kwF@#u(_$>mQOKX1
)Sܴ;6:hZhU.*KBd4ǔNF*@(smwǒKnF~R
Ev5[sJGqf2?&?nv1WTu}Ѳ&7zeĤd-6(#!xZ'lF1ιUY_K	1hI
9KB@R1qQڲĔ-Jk4!$t$EYEc%QA'ʢS*KM	ymQa ė8 	.n]pBdi9/ZiLƯhtps]ȴv
iQ\UAʁiɘ}2(-WJ
:ɈrT9Pڠ\Ĩx`Ri|떵[~Vqnqur
L(v	_v걫\+%vՖ'j__d		N
 
hX'rO-=,2k/R+,b274|%	+#lstt~Z['ktc#8ʖoUݍsJ/Ͳ^?
g׽x)
ͨx÷ϴ<@/~immļsM
XjeM
)x55Q Zтg!oQ&TqggBnTyB7{ޥaoʾ	L'sOѧxQhc]]>[V͑\wlMy-meh6DL[2T
jZ2]NAgT,nqT;UਵªաeDYٮ]n<3Ul;LH˪m!EJﷷShNQ,u$%GLpdH:#n'^\V8~q1:B($sU&*s2xRѨ=)dnٔLIHՙY)y0|`O|3N`8|q}VW7_G*,`4Y^>o2P|C/i@
;>~o;ͻL>q6NYc"6)5
r
u\"'6"ؚ`j5=Npee}}6%dέͻ6z^5<Aϵa<mot{jMx~4€Vyx~%,G騷&-Ys1EӦ!P6n,\7&v(\Ƿ̵C[giCC.)U>u"<*#&>ZKrQO.(xD*t	{/=ӂC~Knǿ[p4R$AAdrxQr99>Mӆ	&S1.B"CHfOMTH	H%-O&Jdf1ui
1J2ٻ߶et~.nsNM
}ڊI#$%Ke
)9\vfv18 Xj֜I-rKSg	kٛ9w{'9p`ߗ'[$Kpb+{Ӯ
~);KefޝIT**kíh'G%-AexsDcl0"yHvOkCu`В85-v-Rs掑sp!ݛ}1
`a[*k+4<CG&F:ZKo$A0t7z}lj;R#@2D'RȣRFe:%ԉFXBcNXu"R*=\?xϣ:n 3/J$l3m+|='>:EGźXnPf,0-WGˣ%$Jd+[Y (w{_Ⰳe9`Yv@-=``A2YBh&	*YUpbp`ĶcG<\hnxkң
E?>PAh?~REhVy?-=
jLֺ6&%YhZ|z܉V!޹/ߤP OuC)"ƈ
AP,,m$UQ)XexaM,
H%T-L%ʬbNDbKĶl5FΖAdZ)'|肅|T)xw{.1ē~ӆ_-7},~9SLD2FQy\A
<OIQSO$"2/7lC\'
	G⃌s%Ĉ[n<6޴ߊ1C!!< o.!_)jl
ԣ|
;pu7S$77:5pFI&xᑛr&EPLl`yW

u<h2;ly5I	G>pA QAFr6NLGG (
FyZA0|zJg\FG/PL`iFq#3VGI(Ak餾5twBAXU
2ц␕RԶewh-=-KmxfKշf
sR#y9OTFOב3-8 h1;?*B$.r"
Z({KLzC6C0Osh\E->{޹Uߒ:B8sɸ?c<<ŗ2I'8]s5B)9bT#jxB{o>)~}y^ǟ3W?&1ǜ3{WюE.QX<{V|Hu,NҀeH+ED~^|%{Ý倘
gB$p_W|1{)gu ń.f9jY2,ON$xOx?\<
0 ;ǵw"&a콪X=1[	[s)Zפ&#d^\d~o_WK	hvIJHf8s+?.\2PMF7\YeD
dqx/]	q7Wx)l
gOx2[OQɿմt+_xkӎ+b!RVZb]>&wVsX"D%/*ilƗVSj@(*Ce<+r
P;S-hbJJJ`EOO.YZ'w>X3X	m40	=pP[OS~@ө6V6mD7^?]?R-8tOKZ.,}~z,bǣɬg.0VVʭ?_ Oď///|-A!{W<_s
nYj":&[0Cwۙ&nft#Yy#==ލY/_>ѿ=&jD*!ܷjT}u
F}[FR,lg21Ӂ홄ƧqG-7k.`<_kfIjo&h[-/cv\1״fS'g]l~q{Nr&
y%jzu?}lJej2dD+*-#0x0T;FIGp`TLgk遅o/xbseA34
2DB-& 
FK"(jŹ!I?GkKT1hRq%{QVxpmz6{ȡ͊]$߷}h\(GuhR<&<
yMwC
6f$y|GLlSL63aƄժ6xPjaW2hLDP@KG/AcGs*˨NYMc=vT``rz}M1Ir$HN`.uuQWz'}+}X=X~ax@x
M"iHypF)`4,u5ZK
&@In7jUto1V;ontCٮFK۵	?ɓyRـ0t̢qy'g'gOrnjAxgx7\|F#9p>/#ך8/56ǗQWarQQG]
JΪ5z	Se=&ᓧ݆jKS$ݧ5r-`,%[Y2%P⒂3?ŕ/۳mz/9xo9HkY㢳6n%
8P4ȣd΃qVrɹOkm}v&n@;VWx,}3WE5c(L+`ZKІitҸ;2BJJ7ۉuW+B=!Z!Bd=IZX* n:t0ި\"\SS{rиX0iŤ32u ?%,M[]5pJs&P˝J%QKIr'T@epNX@K(]A%U(G!\	w@I)"h
{oEZ4GdS'́s]8|G7̖unA`=|~/5=J@OXܠ$6O37&ϡh
rdduG?bD)Ђ
%,d4;orM:40J('hLNpSxG#ز9m>\,sJ-ʟD\x'F*ê"rsCP9ӏ
Dk{^BxP)TQjrQz$5ML9g|J]&b홚E=<ՉΚW%10
gr{9+Αb8֕̆ϑ;IV]ݎw4TY%lm]U_>$.,iR!)+k_DIFКlA7n4N׊4o0mc0q1o9YޤB4`yeGWzv5p,Q=&FmW ܨ֣ndZHMHi,]_`lW|0YlW6vzWEg\Mǿ9:/߿9NOvroi=y(j/"^0ma}gR2?wY,ğWG9띳aweU
=@p6.{iDzEk 8Ckkho5Ъͻ^>}Ƶmrø9.h{eJGMweδ<@8lvӛ]G_#%/GJ`<)ʄGBsSPyc8I;iz`ʸTZǩI04y0Rȕ	yr٢>D\Չ*pg+\f{:t9q/JqMJu:jwߝg73ot<,2KYWA!뺋繅<iNJyJ)Y\Dq72MhdxW}zPσ3`>*&C0I:5,AQ
F+d5*NJMASDIF0tkmR4.=:z&tU|^Ϭ#z.hHmګL%d<UAT,ߴ\SA"ԐH?$VfREvuΓQl9rg=PrUH$d'G|B*t,jae2}5qD7P $j7$|UG@A	m_{LG[xvTe7aa?ԱeUJIוK9V!9.hH<9ҽ5OH;tNs>
T.^1v({ZRL";cњb<[n{

$]]](	PPَج*Goy?iԀؽIzk
>zL8~J":WilriJhBB{~v~B&/j޽Flf/G`}~9}ҖVȨb-
Ɓܥ;H%2o]bw&9NPyT+:2'N>us=W)5
|JsGiQ
#2D0>
iH-l=2
%$|M<(E\ʛ6>_\De*֭:ţp~68O&|LKj@BI^O2*j'I̋9L[GTwst{ogL=֏9Ղ~yq;jtYjccŘR:WAA{c,hٟy6gmtژ18k	*A1>$-'D#"&rk)D49'Zr
ʷ9@!r]䐜$kvEjd`i[oM=8vK~jA1m
b˻O$~(B
ό?u?83m'Yh8qf^Y)>ͫ?Љ6J`u,@{1W93FCx\ԛ
8gI/2ǃ_N5(X<ǣh$4cVK(Gr>KWbiP*j@˳4̩]Y/iETx5\6=
woM~nގ?'E >`٫j&ܿɥί٦{;*.fFv1r[ϯ
xow?n,o/=zeAs1G՟>~d,bX)|GZi,p'w:JcC:*SR0%BtaN !x<i\lܦS2:`0(.FBL\deFsNzyaZ&Ξ;5uqԆaVO?dzk{~}כ
qϳyXanZYr*דy'=G;EHl]
:fHsxԺբqgwtn;8<OZVԲZxo|ΗB{r|(9_>s~2UyY|/W[wT0^O[)y{BRWmr;dn9!OA5цK
ZP2oQ)w3+ɼEZd"y_b2SLKHެ5|^-CC¢bau0*`*H&Dڑ{x8q\q:]
xKgAU;2OŸJ0:sx\o)ebn`d١WEf-%4G㭓xD248e@>m%ݱ8l0~LJ+Q9H2tXW
&xuyP!vҺ}(IrYfK:)9PhK:s$
M2p-3smn7Li`}J	Je)Eny~	r`5vz׎s
Gw}%+vg3,zr*u!T*j>=t1*$*%VjUEurBy:xVs)>7;\bngRTdR1zf/GFҰMHeTcbJ[jC}{H**˜*_n3?.z.?9d1 Y`Jk *B`2*.2zAANdeK/bB-H,Jg0,:\8#%8 3#gH|Χ?>׋Rd+
C˲|nwSsuK572+M rL0F7)GI
 peL$-w0Z"zl=N%)Ig	9譽)i{cvO?]mvL?O׿b^?x8&}׏.~Isj|w1	We^)r6Be5l&
~^~`Lp=
~9Kte1Ӌ!^#+
`ǻ.퇋!ZE
 Ug(JP45(LI3UףxJQWr'TM1]{0Zzz9<>*-wFro+
ik?(D4+] 	?L{IӢXp3O.CN/\8eJR[?M(묰3ګEQeW>.2I^
WQ} (%͵!o/4>/zZͣ\[ND=\]y*嫉Xl[v:r~*sY/UW^^77,T02.Ʝj/n]K1"yutց7 xv;?H`2f1XXPK%(
ƜSIR倢6N^N1L3(5fGAnU\NYs(U)knٓQիZtU}U٥M95Źp:lo?j|}wT*kԨuP2UB2-ُřjU݌ƥ>T;eQ%-,TY7Zs!+>-ݟFC%ƣFs Ǹ\A
@1P=R!Db:QT4KYJ:FgRdYGͺ;M{9am6&ȩqP#Z20'ddDV$t$䂆n@msmOdd8.i-	4H4
n&&tP;#
'b^e9c]	qOz`2i-%u.׷:=䤲ޠwf稞Tߍn9N}xIw*@J1
8:
D]@fRDԑ)0f|*2:Ř'r`:&YtMJi*(%R]Kȹ[2vt&;MeX,zy4k-9I8Lɟ5(]?'_6ZhBB}I1%\*ό)_מ+לTbsTNC{.2½geɕ(6TNk#aˉ	ª c9wKl?+R3-R{x79GfG&R1Y43D9W܉@('m$vI"dQ8h5HP,V8$#A%ˀ)-;#nyXE6]ǦQw,^"%*w>X%-0J%cA:;$i82a!0ݖ'@lRpjҔG-XrGj&DLh4!;u,;#g8\y&.:;MŽ\|PM1x"8^󤑑
*Bh] DZ/FDQAN(5rqrqgĮcCyщ<4-;2ϲW{'q	Gp{?>V4|{}?>F-d[U2cx)w$WxajmD>t`F'6JQ#pEjG@ZYpДFj3"'EJ8.p $	IH$BFQk	y?Ju]tE
M)K60?`|X ~U0U'4	R`K),e[Fث:["<:KXt
Fk'TuJ%21.m/c* Izv.I!$TڊŽ2xh}NNl6![Bo+{qTLJѶbNq]n\:]e,`m)ХHVJRj`0iDϽW?wP4+TSD:yБ 
L,Bˇ
ц<E2ݞZo#E|νw5زXm0	tZՈnq`Gi)Es_<5Z"puH6,KQ\8\lQPL`&Z3&U
kȹ{/ճu<'A^{kWk,3v\G,Q+9^typiPT#'768dAE`j<"<%PYXΥΨs\tfJzrԷ0m"my uWÓÖr7##ړJ4D%uK}"6Jk볥"q܉q+x/΍m9G#CL0%OAhô:i܈H
p#)(e^ZQNb(dZwjJ؋u'HʨA㍒+MmNEsA6c\cFYL:#ӞǗߚjkZg!`h:@rgR	BR{	5AyC_j,
Ri4裐@JH; qJ}h $˓0)OD۫fӌw,̫໣0)~Yr~0!Nz8ՔbĞtpKn~~[A7yEmɭFB,T~Ǚ@J*8|ƹ28@6FsnAvpANpc8˃mYk~O W`q7)i;9-Ws`)/IU%|r0x!iI֞\y[77Ydh#|̳wLyobp^Yk3ȶ^jPo^ټKdG!?z:b,6-3~\Pc^}=8@57%圄
e\V#]coel2Eg8C;a^GX+8Y޸#Z'{b,!@{pe
kNעQN!DI81]>H~P$Tcl	yn6EnU}	r<ʧȬ6pһXQ?{6l
Oӏꗀ`ln0~	~:(G<~)Jm=lŘ	 d<ͮ>]]]eTHsH=0#=_qz'MhpF=b$.5u2jR14ݚN#k:89;=VASY
"=V9]9pWJWkή?z"ϟ#%|'x_py'!Z^儤{|~q7l
%,	
$.zdժU8s_+?'ГK{@r]e=.z
gupwl>5NJjWx9=T]{WULf
5;[LC65`<%>xb{6].r!Ëf<hx~QoOmBl|@)K!'C4V2aN-"Dm%1 {rl=<қ伊gA%b#[Y˵C6D̢e Θ0'-
*k:7zSp6u7BP`Ff$Ƃq;WWD+mczΘ{c	Y߫1ƘC8صŘC|AW>PcQΘ{#-XUh[**4HW/P\ȷU6(zŋm$QcD^ś/dZ4D9׈@AuHoa.޺hֳIZ{1Ǹ8u9B+x3qMn>6d>Mdw
fÉ=!ՙ>*BHyz7>,x?VcV,xsWq+3JoȩEy'5.A_	:KFc}>ک-SDΤR$J4u.QCVq/~w˞-[w)5'
%rVЄ{,,@Ge$!9H)e/Xv:8.DP]7L,h5T"=&bsK'hs5([?q+t2+~/釓g<%\m(h[0ybB(<	6	0b(2&v(IE~*{Y"uh;sT8J;,1P	)'P1@CN梡&(ő@p&eUb%z014g31\pYᔇgܤc]?<η[7sXW#³e5༜#wDw2HSt[mddf"v,'qR"C$:ixQJ{c-#"rQH 
ۼJ%@6J#7
:iyn}?+s8=0s'ޢ"0AS&%W<0mT0hC)a|ztt^9ujVhBq&PwRNfX0;lH_jsݼ(C!N*P4#̛qIt(QUyHɣH<"i1';;oX;]	xk45+*Cў6EfA_gӀw22)cg	͛wZzh	J(KQPEooS3Bgcyr@,|xwvvϋ{ڒK+)	eBz|ě+~<ζI/ʊ 8Uշ|DW\qGZ	U7bE4w@oNYb4G
OXCN? q4;<[(~qq]>:^]CL?;Uɯ7fǹ7^#FQJfǾ=́'Clf?'hX3?
B=-1	0Qfx~^_?ŝe)IYngYFNTx$+JzOE5
iJ2:	łA2᭶2t5L̻H}E
-5&#FL+XJ
LQ$B@bdpA]!pё^#%Чr'gc1h{fƒC^of(m7瓺]Zh[NZ5Җ,s5TD¾R`͆`O?4"w )Rx騷%g%.q%0	2C%@֕kW[m[(O-CjoRŽmf.74!h8HKU60)yH9&'<˃8Cd@g({Z Hk*3(ʛxˌӹ}H@22NLYZͳJf)Sނ1%DHA/Ca){dBAԯ"=%ɦrɈh]YdS&덉)cY(#t(X$q>MUf*uE\_l򷼘o,Bϸꔇ78|k~ڹz1`?aY=zZ9דs?K&j3&!NB=?0	7ӋPB7y͖S||>ˎ@ݐj=nyr%NޑQ
Tl?׬ZïO*jV8~ אq'i|n	޵3yhi[ٌ/}iy>SɕoH{Gm
.>\nh41Vl8{MYNӷo:zy~}XGjz3ӧݓukmS{ᇛWaЀ>*w;[#lI%6{fp/Z:[BmlO׹r.[#XG◳O끞߶90]lUݭ;jW_-Fiu^ZGj篗ױ>f'eR)r~PPs*_\<ӫsbsW~?~Ҿ}߯޾^q#ȑ~!nDhi¶gJU6LF5>ڭ~kU֕,M+\|WBϚ۫ݿi+j[7ާigWOg;v-ҭ)RkB|Sm0O{M/`AoGxV:ڄOK1`]H+K@Z=?Ho6KjC^8}E
edFIU	+ P֚W:8ɐtE7tz9$
ym=)ᩑAlʝxuyVvslGli	
Ю.6s<3NM(9|ضd_8/x{;w2//}:}P^Lfmg\Gtt tĶ67hzgFb}m&::HkUMA6KA"|iH/R.aHLX&.$ <[i}4IH[2
"z:9*s1TSSt1Zc"Bge_I*}g{&;wvigWDЃ|((QPË}},?^ORCvmA=@r6ZFݸ\bE6"Pyh_,{+z@yjpJQvy-\tIZv@-eqbD(4^P [掅;k;6Me}i{[~>`trwrs1BCAL
lѐWq<z5t*%;$GFVh'ysSp2\FSKCμN#
D$,cǤ5Jl%dX'4kJ“սjpt%8gͮ-؜@(6J[")1YJg
M5A
tKm~]jdA@YRsg4')%+M%{ExX-z CHE` &}iL]AfJQgtۛ8{d/%ihv7VQ+m]L}]*I>u@:3@#@.=z1#^=^ZK|:e>SŲhCm6Nly+G25 pXoR"+aS3#P!e0&{HѨB/Ȗi5g͏zXcYPIƒIU!B8~	"ɚC&5`-DG۬B,%u)+gmJ6RpW+Ho|5,!$
^hC<T!btjA .;EGr`FeyeaǷW1;c-Ɔaeš+*Lnlvz1:/O
MKtuy{ Cc:Fo}6c{fܘ;ok3M47d
:U9u胐%B!4Efh%RuڏI_žٝڦ?}~MʍO>N瓛YଦUczmZ'?u.U<-02е؎H@H˒Sn|o>҅(Ƌ)5ol5)8©m~(1"%*	N6L3.l
gP@	4y@FR>gEZas
THчP3;*J_w33dzgK-ןڽxlOi\o#7={4_Mc3+~\-3/gec@l˯>-332zI}@uPʮ_WiLz	uh<7?on/3;DWұҢz*wXf"Rϓz{}CuyOIMB0M(Y%=>Fh#DՎZDaIx/:kcH!m|
L 9c_b(K^Ղ')rA;	DI9[[)	y8ڠL}NfKtW8Ќ"m4NWyTlXӬ]A7ۦSUtHT3lU|ǗƉǿ2Hp2zdwH#-/SiW*nzГhi"\4rԾL7kT
W'+6N+;
\Yp5:v\
Sĕ6j0
\?ju TZ$q1uށz=D/H^2T؅7>H&4{㧡4c(HPl>AJX<`?O

hPiyIJ3SJDE}&4~\iPpu
þFEZTB;tky?g Nд}Uy}۠sߝ7@ӾFd]97~}{
fC/7C:?WUF?f|v5wwW.,w/̢,w7M{lc׳yާrvwxguSUoX?OSO|l޽ݟ~vݽ6j_LּH?#?_Go܋5~\QE M;	ݲg3pE
~	}qkL`~Ac8{g[ۯ):nm{˶oߕA?^N_ؒeZsV7j6ϑ3Qv5/Ga˳VƉ^nj/gH//Z}n?Ի~ =Q1&R5rӞts09&quB אDkDo
ד_k攙9Sc.T3)jKu;`O
cGIC#;iAV0cn#ŊddS7B&Vڜbd?Zj
jhVU˵`0"17hFR)%߫]n84˻7o~zK/z)5[CCdWۥA2pRGiĩcĜDhsRm$T/n'D3!၈&3G{VV4j@@1+҃:Dc;
1iWͥa1ĥڻ^bGF[j(qxz#M'BM]?OFBCK{9ל9>ICjssP0-wZ"ΙTRU.r=RNP ϐPca$o(5I[4a##[$3%H/&SQZ
Q)\@55qdDH,!Xs|mVb s
05{b>'wмX\)NԐ
7RN=q/HfnJ!(2>սb
:=KE
Ȏў6d#oG1YeOhBnXAC向[`jbo
s8o]8Ѹp0P8QN[g_GWA[<;*akBhcnu/yXLkzPDžĕ&(vm\sQnˮi`T! 5ٰ֞7FbUh[EJٻV8V*
>'[>bpHnAJRq
j VjBl@ۨ 	
	W2C#UBŹ*R}3iYphu+j,:MH&{ŒSHuu
'TV

u+0`Q {()HX "".ZrHcggbM))"FZ>G0rx`B9sRQ8(=[,묏<`mB@U+c][}FФFu 0(cZݭڻ^̈KSM+f#c
LӉ1$Szu&+?]^XܥgA>C-x
`cq\] AJ6]#G3P
v2%,FnMŎkI#^DHY=x4vM`H yB_
7fĞlc
&%
xKd}pRSP/O(7"fhSq.(;D+	+Tp=eQ2mYZ=ZS|B;lŰ"B?YЊbD,Vԣgr9Y@l,2BϭH
9AG=
|,55DqI<(2t
M5Fi\mO`ײ￳Fr;N>NZAs8%U8L,~
RbF8HLZ0^4= kSqpX#9>dp{Qt~elxW0@۲Vv
㗖@oBwh5sp__q;|o*q%mljx0H2K!@x$6)""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ":]ME-YZOJm:Eƈ"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""tI Q-wA|1$l$THS$ H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "3HޓgZUG'z	#W'JR+H^*;dH	t9O$@OERHo|$G}[C$@D	D$@D	D$@D	D$@D	D$@D	D$@D	D$@D	D$@D	D$@D	D$@D	4TH|i=w.ѳ_gK.9z(ίfG@ '}Σ%,}9ŀKָK1N\dɂ
{-+k+.qY
UdWgWB*LAv.ƮP:~]J+) Z\J+Tk
UZ+/ͭ:,f }t+'7Mwfǣ;S⺄fPTp~c&G__h1_Mp7P9芋gcܜoJu~r+y
i%/ƦQ245v6Gd'hV]i%,'	FʕbW֘7)%lAvM9vrtG+&:AZVRtU9 }dQvj1JZqv4$b6FZ
ݮPegU֛ؕ`l1vrE1P]JMsWߏ]{V:cZ+s'ٮUGX`?^ˮTRdWz|vWk})vj5C+T)
	ڕ0>%ةr+M)vj]JȮNЮQ=]I5DWֻ䝧-dW'dW^j/W0;[)/bw^&Ӷmezzsܴe3hMS?
P/3pg漻B?n.RǞU߯g+%JOf6_@8"Ƿ
ɋ(zr>ۂ@~T{ntT՛ּy\:ji	sqs^UQ?+W2^n3jp˛YA
a|;6lmS~=o~}gJ$PI&$|yTuNJ|jY7Y]US|UWݔmveYFZ爸Y\URʺRʇTФ(YYN5Mjv(d汮fQ)RVUtN=Cqƕ4Q,\WubmP68
'b
:V]ZXJz
/ǮP|v*
':Au7]`D
z5tk+Һ Zb
*Q]Z;dU:Zsv溤
Ƌ+m)vjATicWUϘsB`?DW?֮@gBܔbWVT]9.=+i
eph痣?ӳњ=x,\NZ`](]붂/E{EbUɼu}ֺsuY
+@j˄5>p}L)'o*@+Ja/z__
p=n?W"}j֬Y'SIk<ooS?O;0¿,;Lh}_ۣЍv|9Om}5Ώ[7X^a:gNiO+|0KjmX ėygWۯ>>3M.?W;>}_cޱOZw
m/r{L`d||~bmjc:HMmY1W7ʉ1%Ǭ5n~Zƙ;qnpsX*ypq2JX
a\idpJd.1ac1Yg<>gv\lyFh߰mHcmz5HG^N?E_ϧW⵻-~2.viáty~\oD|^
uמsC]LUZ]lGv2GHmlsq{qdUculY%ƫrr5rP8G笍:&iQMnr2`:hvu&{΍,Up\`'n\Ɂ[+FTMd=`r_~W-bz/W/^^qθ0g.>{&_ܗJo۪[;Ow_=u6qÇvl;Zy\nbjuMxPb|-%75II1N@2%#$.".1)3uAgy\%,8%k磪!u(82<$ϒ:pLz09L,wn:Pt++;3dY<_ du&=8N9=~u1^ߜ節`ٝ>y`,3lsIM.x-u 3Lk}3tY;@❶쒝OsmANjMjDe*@TbA9]6RcܝB\6(ho,WmC:fi12jg>`y^fÌWqBH1>8P?Jͺihz>I8)-tpH228nLc֬	n P6Z%[#ԑۑX=/Mfa^4hXMZ%Cf9hR.S]xb3ʌ5M	"n?l&,6,S3uo1Ƕ1xy5zne:a]Ks9+ͩ$$sloe'p4HHmMh	(aKf@$Ld~ޱ7
uߚou%[	X̛mt&G-8d	"đW(7OhAO|>=eQ}>u!P2B%ѹ8`RQ,ώhOgiU%y͒z^6	vq]d>q]QlntOңEZ..?dM`v0h@9Nbl
h2*ɇ<փ;_ZBڒ[:֌ڌݘכٜ,4Gbŧ@.<=;U)vV1nVgսO5RF>^w|hZFUmM}RU9EeQ:=?VG~_~ރt8SMh["ViZ¾gJϟ&oT]i銮KuƦi	:Q^}{NFJ*saE=uh4jZ9OӺeOe:$^8O-[?x/3ߴPfWm[Ϛi]Ee#*F"g/:Ēї1"S|:KD#6S;VʥWϷCt=Anyl(		|هx$KuyR!J_ÙN3??p	34WbO/|7N`m+gsw|΁\t
V
a;#$Tz1\9re•*)+J+~D*pIWСSa^*.uZjF
w!Y8,Vއ^ONKIIdIyRg&%9Y鄎B
:fAA9$ @ajɿ}6eNře*DuĹ۬vc1:.Oydv'IL4Djݘvzss1m9kuaOE&Š7tNu:e'
{!
SC"|2Hc@B
>f8҅,k[mp\&7D
dMHʢ ƪ}`.FkLR$sbWZc*3js7$	 .x7y>
A*ځf<'/."}+P}	(eu{P̎R#,`c#+Vl_wJ"y_4@xw =P]Yg `QEҎ,e &b(	
d
]ЯzNYWbo}̝<c1=;gqELq1iq/܌ƦNlb"	1֡p#Ls8H85b3ALL(/ZaUH2s TR&S
)mEL_9.À'pMak1	W)G)(Ufqˇ.:2Fo<ޝLcjQX:o&wP]0ubLxxuq|IdzF}f-R]bs2]t墌R8=.ȃy0pm96xL"K&-t9y M``&:QtmqgA$a >dWG!e%x8pڠ̌)Kh}-Q+qƝ4}=L0~|W+ylQqDbCk$H,j&w,%[ʐ؎&V+kZUw`LV(,DY:([,t$͕jyI[9]DUԖC6D
RR{0XV;{+)Xo"-*K(I*ֱղR-3v;5vAE~{~=oA[۝A;܁yg
*R2CR:7dhj8xۃFb]KE]QaJe,(4Eb[B.G\x5 a
uAaE)`0,֦z8Ρ°L%{6WK	緞*YΛdEYtلy`t!8Z$!X+'
\.;03H3qw=׋'@:P*ͨT( ڐB!Kˣ,v>uX/%VY,-v^\u3̪2B?SIP>=ķ_cΧJHl#+9O+wgsg:Gp|4|>LxOl¹b!K'/-8})ƮHV"wޢs2 y9rNh;n~wZF_N'O~YN7vkʻxUʝUr%xOё_{Sჲmk	.x8
=Ҩ`
d4ru_?SFA/Ϳ&'/-~ۭCw_3t	n 9?񊭵3onbhh6˸m%L[O
PxQwPUmlW8;wnzHe`R@eNN5|ȥPrZGuDZªPրDYx|@f; gBZ>Ә/~t> R슶@s).Ԩ$Ё"@)B9X?MD:h`YeS./DgY5@۞FZ@dH'\trԠbN!XJ mU2WaQh&ΞA'󋯽/ny䴋wa^{T0u6
{gB_gŏmUO})ȕr|o;Ԋ vNi
J΀#y܇]jW@C'UJVLDaBA!@VH.ktLܰhEfzhȊ*VEDPSNJ!7X4nu\\1RM>wjDTϓ5y7aj
.&b=ނo4ţ7"f1,G݂s.I>CCi:]:"F2&-%7:*z^m|~()L95	U:x }֙xM˜`Py:CW83x w;Op[.=qyw-mPqn6Z;|1\U\o_
"K:brrZq;*6>4c1>IZUM:}`I
ÑcЂe]% eĤ`5/]d2P9cU*AB:I!e:XNHH?x1~9Q[ت1됣N-JrNŁFJ:+)XUȄJC*bӅJZX/7aqGeV"	ȋIJPl¥8w/TPͨh}T[ɽF9pHI{S27Ĺ[c$o͎C8hXo	bv
ZFWݥe.UIa!EBfd(
()k2d$IGQA`8wި_Vx(F54wW `9%
ၭt	lIam!lj*[-K3142`ؒڪ`AX#6g<WY/Vl6KݠH1E+ϫUDvC}hlLDmQ
^|x08츧>܌h#J/U
)XK0ydEhڿ<
` trKJٻ޶qeWũ{mϢŶMQP"ulײv;#?pjN,T!53d,+%E8==XЍCfgq=@f>D3C9SXˆ3P伀a`2J
8b*Qg$
##.Oո!@\E7$/2o~^P4zY[ǃYFeF
\7U/Y@ֈ)Q%XX݊PyW#4h}iFFbc+ebŘ]؅Џ@mdze~pv wd/qO]nLϻXf<'!#.F͹oYx@Z,SS
oNwqMۙ.0]5M[?s7U3d^ƣ*urz`r?lwz0^Fɟ;o_mcin
mܤ:P`,`wɛ";^;;$*$@$O>MKr
RN~cL+Fn;y>`[gsQYq7 V?
D_Oz?j~d4->@Bw/Nw4+,;(׼ҩ
ه_;.1g ibSԫwGw3pn*uNq4TnnŨndv"7ڳko$mOf8_|]^=>@$0XEu"`IfQ+_/'7CA\w
W;>TK߿RB{RxZt2ܤ
Uj?n&>sE"MԼESS[A~6.aq28)3wcqya;tX\J6s~blW'D"aLoBԙl~zt/y+*G8;>^5to@{QheGTr?}OǽΨgL[:!L?8r{-)`عXkki{4
E3蠵^#|b^
i875Yų*6֝n,m
 ]4jb:>,j>ݺH$[!VyN
$ǏjTpf20T;F)$AA̝q75;t/k]>}bہO3hYȵ><2`B-&
 F,4x@QlHo'iFA]"wgݭ]X{sx5=B	ì'5Qք+bLlfޅ~/eUgBQK]*iFsr+VM\U3u_'?}&?PJEl}{{Lf_HNHEjD.E~0:]豷&˸P"@)}aYdNa
e,x)x|fsg.f&\kAf:6Yi	6$S|hNXNEV_@)%oN=S|B
K+M,th;]!JM:@Zi#+lD4tp-%t(lꧡ+Mߺf{=fhj#fTkAWf2
]UP&IDt-.g4Bԝ
] ]1ÕDECWX
rRwPƴVh
]qu+Daҕ/v/Tn[ I@s;yoy-A-0k
|vizP]jd&DiSQn00q~9唟cZC*
.h8۵ҽ*8q[oИ#{|tڠPhU&)rՠ̓_HM]V/?]+d+q*mqn#
\gB$M#ٝH\ʲHG@E4M ,,l%eIH^*`.d뵤yTK'>Z479,Nna@#$hk[47o- fk Ab+,.'t(mJYA
}e֝%mJXX.! a>{PIzvKǣUqYsϠT:	N8Ju>CuTB2	^0u6(+?=̽$/*awT.(9
+&I?Qr!GG7>q4TiѷUڲ#UAVhoc=1}N(8^hsRĪŷ?(@]@]I&,-U(iCWHWilDtu<wߍ*NWY$]iNOWX3
]!\!b+DL
Q!ҕZ]H+B*O֐uute& }E6
*V+D)IJ	YGXE%bϋMSXɍ>iWjcfgo:-պt5+]DO-]!`k+*!:B՝%g
] ]1bFECW]Z?7DnKb$%:BLBWVҺTC+7L)lꞭ
*^w;4蚢-9c6Dfp7Ckd]mͺmꮢZZM"+!&BTBWӺ+1YW
]!\b+,@U
Qr7(]`J㱮.'t(UCWIWȥ?˄6m(3řzeFh"ilXaj	7+Gpet
a}'ҿ/.F=!r$Bh
\jrB@o?9zF.owI@}gԒPet5AG[jRG=8J(B~6	ݮ;X'}?'5/j/^t9C`<4)WU
'_}o#][?nW*\us[9θ;}ϻ76W~pF٭3
3~gG5Won܁pDi^xN$0zܫph'<13Iؿ|}7AfwU~Y=J#pruUYX~]pf\Wԇ~>H/D0hgI1˸Y\0-d!B`Zp-SPfsgx\A=p7P~ۄJ{nIWz'}Ѯr2nYW-zpx-:`qa(Y<_,s
y%-=եyK.5wh)Q^j#|Ec6˺X0h#Z9,˛&gݹB8nIÇݠ~7Ž=dQ]2/,˘)#n4YR\]L:\&,y/G/_PZTeEURtp3Ӻ+Y5|>MXINЙtdbBu'C?oϓnWßCvZZ!8:9ֽYTt6
1eEڶZp\^UǣXѨ[.vk&k;̕
TxZ6M95UKaXmܴ͂ϸgJlon3?[,a`[J@%̍B9jz1SM(RCkmH7@Fwuȇ`{0Qm1Hu%!QM'F$jf83Sujc
p&>ө4+}&z0BVl7r~3"%'VAfMZx1{ڤW'T9s~$s࠲41%YTh+%)4b)nV+X9
axDག,u[	ձ	
~2MIjA;rc8ς೧fO]orhr2db<L\f5C"ՌlV 7h;svH+`MW!]ZjYbo7ϤoK.-tS8::b_3F.AzzI<~С-qFv1y
XfLJyQTB霉er]և$ul{8>
."bNonקA (7^ .Oq؛q˼>!LGoCn5grYi<>஛K!Cֿg?
u\̲J\tӤM%Smݣ3Q=Ҝ8+f8M`Ҹg@팵N&
f2+X%q]{OI9\u~ʜvp99RR?-:*/	%1"@RsL1(j=YXf1qRr9MsNz;c
Vk=j6B{	ǃ@ejGG-9<[D-Z@JcTI:YYB-9Ժ*̙
y@vu>k^>%#@NZM)0+;jlwT|W%5c8n)q3|Uft
VICwDyʣQh֫dvMXO9߷LEl&^3>qwKI;9ɨ1Gi!"SȃGdyfЫG8
aC]
c,rDI,E0h5HYyds|&՚qjXX2vB]U{«{^ff{
6?|I&A2IIʘc&ⵍ`mHpƱd mWElIHc(b'dB@$0eFH;ɬB9T䈙Q+#vlG0%y(]:udot c'ݦIr%;!@XO,sI&/*2.Y"cU<ڀjH!3$1Jš,FCđ06NHE
Ώjَ_`<Df""TFDCMKxHVpFMeI˞)@IPJuÐD 
h*":3HVlJ3!h4$IIW$UFj͹G)ҫf\f%p'e10% 3!["BDèiLTR褸P&8flJ$ׁ[r`_oaH@wަbtTtk#6f^'kMho3ӔzX6 $m%o*]oac]0
O!ϗ<`%TFeL, yBpD?TAB}wT3#Njz%Y ЁIh	2	 9Rqh#	u^_ӯq֛bޗ mkԃHQmeHFVK5@8dKW)ॅe`A$<\$3F	D0S?wɤn0w&RFmϣ˓3irB_b9ilh|L{<8zbTy]6RHk@XJ齞e)v`{_O~m{mDE?	eTpH
]%MH?"(.pz1.ζJFg%Ge	\j'hozgTH09L^s|?.-=}hbb 8CԨYR<黄_'_,w4#?^ldB+[ˎ#u9[lmɣ,~㷽|4Xo0Q&'Km?]7/ZFjCdZl~߳c_2`iOM^ﴤ`&lnyA=B hfcXe
]wKͷtKS<;;fhNeu%YJ&NKW|=;f޾^n)z2K[I|u6JھQl-Q;v7{Mz+Evn|ɻYoeP>Ɏ͇߬iY{j|@a'-UGTlK4'&5N-L	sλEsٺtaɖsv5.dSI.MgG8p|5YUzz"k榏mkpVճq'tF޹{WƑ{HF`dNn,!)q-IV_u/ILOwU:2%򴽞6/sxQ͈ƘCӅБۨ49N0>iQ
Ήs8>VAFxRYEe`.0si>wvތ-Ջ]RTA?{o>!R[R˶fHk3rinfy)a`t׳/]&k:'jm:j[_F6ouaGRwKU_C5UbWFkw+*&~n*Kau7߼6|?߾;D˛woO0P8jsss^{Fi
7vz!|@T饓0ew7ܴց[?6Hw뛿o޴B]5
͚m4ߤ]e[v+q>Eާ񜏴|\GXoKk壛hA2ޭ2y2RP	=h< g51Uzfq"r
H+f̥λ}Ty/o$JÌqQzmmI5)ckIDG$tNg{u&݋0ety}>ضsSr)ϭN'S<xYIG1s]IJcF1tY[8SqO3>&bM)UUc~5wW)iX=\F[)9?`j#'f;j-?&hlGXP	EL 
V*0üyw9N|^4՗S0K*xڷ4Lݪ:-zmux χ˙m-YyNGYQ5LxquMm5̯ը?n	4Pi^:J+({[6{$>C5D2eS!cFY;
%{]=`Y

(vXs_cs/^^C:?ёo׹?e	`f g«#}WI :[G&}p0r 8g.H\%3>Լk[̱ܸԛWW%Ty߮tU23jlP:=yoljͤyeH8uZh*k6kUn
ad&#QHYu2Lwڔ;``-#chnDt;#gËfSo;[7ԷxC-lIԾ%W/O_\g{}7^]e#{p'q#7;<\c8$WóT >Ɏ$auZkUk.^B/oйFۿ_!8dKo55B	P=ɰ6X]PC}~¶)׹4&>DdMo{H+p@EތLğÑ$)[Xbt+lQRb\(y|/ɩq&7uE&i	հ*(^oxE[rExAih߯Ǔp1BZgii6zi/D%m4٧$z:AோGKw-t
sS)zJe_;mtn^S7j^[R:fok}d,΂K6S4"SRozMҜuVtxC?eR-8dZ)6!'_]m3iӄ&c",q󎬆5NZQ{#huWx|qnd$/Y`9jFq˄aa&"	-CQ{0rHR\%HMLc1> 8NxMwꌜ_Ȃ d~f`ôl8SrH=By%
D1es"=j%Nn/<]RvSҳȹI﹠11NHF$Il6pDB#Ǯ@*pdSѸL98UGRhL$xX;!T>z_6\2E,R9gwh}eD!NI양=c$Sk/XݷahkɲZTԁϤ ݨSEGj@'4jT	
u45a{AEQ9H#@(6l7xCA.2LIeݡjcMi+R'.lbB	aep80k-upLr9MU
;-Q1ʃ^*m4U}*bS帓Fյhg*gc?]t백iʚ'61z#ڣ@gT)R<פNj:hQkcM&zGG1pʅ4paU*({QeW@-~Z>lz&2&%>\E%(P+J]2a+ s}0*+ *Q󺰫CgWdé'GQ%ճ]=\)]=Z~v0*/vEHaWzx%îD-Cήv`ADî

Jj
Ȯ(lT#XaWaW]k2gW@%_؅] v\+KיˆXR5vztoAӛNRCpzh??
&MQy3ט԰,1I@F	"4_|~3DXݽ.rGQ^|_{;hƦxG@kV>0ܼ3錤T6LSW9Fәez!65Z\J!ľW))np05sHז[
˖2mf%ZG&p>QV5yZ[LMxol5egMn	Ł>B"MvjUJ)j8B,v,M;.]SφjhF,e&Ur:=O68KrMRT/e'=FGP-sxZX`0ǿ}=n%Λ?ClOXvHso1of!"
6iA&)sRY"fjKA$\H9Gpflŀ5-qMK\״5-qMK״5-5-qMK\״5-qMK\״V!9M!R){8΋7*Y1JY0%s	\0%s	\$$@].a*aK悲Ja.HkHKhź7]m@-]lλ	>9o޵{bHZݍokߩ x$
)[XIDk1h`tm9u$clW
čp!>E5R5*	G
,W.L_ T~n0~_(ny0j<:JwEpt8'IGJOg#é@sT-}CITTZkUlk.^B/oйFۿ_8
W|Q}V5Ԑ!{JCk'pn&՝YՒp_}!hFFoFE&H{PՃz	{#@6://`FL8
FF`j6);;on3?~ftA6PxF9zu֬{5,c3l{xƃWzIL|~%ڰ:wn謇&5N~P=bVVMt
sS)zJeګ=~4sPtn^SOf6y07{XxM<0llA1Òn(z]|iκ?+jL}ӭj"wڥZ%V%ܔ;$۲hڞIӦ]=ЩMVx#s"kmVSk=|hӞ@EӞ~h
F\INp%cmE^dw;9X,MV{R4UTɦ,dB[ak9	;߉O"<x#K̓|1P1CN"2>M\',0Fa}[ge,2 'IpA9g9!Z44YeHV{Lu9YvdLǣˈO6GNjr%d":͒@2s(PeN9k8`赘
KJw$82e]
RF
g.i38Y21Y6F-@|@RH٤78.TkmPlٮ9jf"2YGNrDSfB		L>nQ\))cq
JZl1cDrM92-PV	!%hXdۙ!>rYnC*lہOh4zq%wI
}TG:3s1Zee&ɤwepI%$Ιke1nt΂̩rh'6}
e[k^}gg@f_]ߕմmTlzB!gDk|r7wׅ}71;m,VD ',sIƫM^(Ud\Dja|VsE2iEq&!FѐHE|HY'FJE	ΏykxF	0v[]ZFDG{c9J3xHVpFJIߚ=S\zş.k+EC&2ٰZEDg9cƨ.2LH!Z
6F遫Xq[F٢)ҫu.u%q)T)r21%EG0*`LTYPgȄ=.;';G+xc <u\D=?kaܾY#r_RhH+D<ˡwz4woi#BsӼ};{Nsо2JF0kU&JɨN(H_"8ʲ2dLe-TrsبDq:U%!Ϥ;\8#Dn3Y3s-[
t|6}^:ֹwZ>ƣ`ޯ̉vsH+lUscYΦeAn%Bbt315x`2Y(I3D`P="
/ظ$d5A'ŭe혱)	,&r`m8L,T깒KLi).p R(E(Zy&!.!w3WfN|<ݢ`6($>J0H`TVK4dj#J#O+SuDžOn[q^I=i@$4ψgiuYqhH#	Rߚw0oEn8e!K)j۲	\W۞Ȗ6@GЍUyȏ}J/(J$:%$)1uH%!"aa|״NP{N:3Q4'Xԩ菿4p4!yY8+\Po?WKUtW
lD[K![b@b?-~
e]0&d4^/pk85㋋Qx1eTp_I
'L
BE]I9;wZuBۛ0Od8X|)w^#``&PC*ˌz2hȘg:WǼˠ%N.=rG~v \r3
#Ȭ#re>GP,My#Q[U	g8Ks;^kix	W~W>F>D=Lv>)4jFFT&.!)ՌxV 7h{mExW[dW~Mm7l6pϟ45Z4wj;޾y8[o֬8gEĵ`$:"iTq\
q9>f{fb\H~O/lKSlz`7jtwꂔ/\)nˑ&ѓ:gkOE,̄(D#$)*x"&0!99ڮ9[\igK\Iۈ(bEK
|z1p\p.\ɀoSAxF6f-:ƤJRQ"ɐIsz2df%rEC)i)7x.Tp:	3)
0(Q"dثlґbΉrN,2X }f(bػRD5{ja*eeh5%p7It0}8V&_>ҴUHT7 Hȝ1&X"dR	:zi,E.zAjWvEmAnjZKTV#CJS5HHCWXZo͛L/|p
.|}qP܇4yH^~k~ظ@r1c?an58Do5V8R#0vDhYPuegN@pK@ِaxs<w]m
>E\լEJD&׀w4ox\o	NN8Yy9>ax?ݩnpk[ԯjg燛*%YRĭZQSWob\:qZJ)nd
ƴj=X6/[kgEη/./7;цKb$G'˹jn76Gdgrrr8ykIƞJNz:kF4v|~Mfy%dZ`ŲO'w뉞_9:-pث`dSMc7^e_2R:>by<(q`mj'~^>$`*qt~J/_~/o1/_}8X`Ei\Z_P[xȤ,~1I_汜j[[e
T$
+4ڄ)qƉq;xwmqHW?wy .1Fcy1`2X<=݋=*]RTTUIF8d5w~k>%zo\g_78^BIy_W	^W߿PjV}tM|F;;IUhBFyKA{uUz`%]Mhsހw !5wkeZE8i-jX%0(IZO]HDWCtM"s:;|1x>pҒ/A:;/&gx~g/ye;H"v6:G|r|RG_quM|^WK~x5j1(zEZ78w>g>|8]?~qScznʨ:?75vaX8Q7	^ûޛͣ:#o0qҕ21/̍OSOcƤדemTmTϷ*2^5Zb*_9wbL-%Cj|ɩsmQY׬S&5fl"oEEj0Z**k҅eb9 }%[2)l;dۋ,@<t[0.s$؄d~<s:#<}KĈH\J~ltҬn2#%S/DTQF*kbU̢Fq)W7kDٺxɭά
z Lj%֦:ߗޜ9wS(Ο|,X$>D}8zqS?^uXap:y%h|zˤMCh=yy¢ECjz.[g$-ŘɈZV8GbLLu	!f¶G,p!Ph謬-bӘyj]o=r].Gn_4^+Sg)]~}.ob65.nI/Kpt"2q8Q/if3nqʼ{;J?SѠ!tBIƮ-V/w7Rk21Y0/Ӳ-3DYH>o?uw{Onk3N5:yryv8j\YʽJ>b.:cR5Cn_M"&XDSNMVT
jOť޲9.gU/&/swoY_ARګخ)V]Wϣ,a~u7NrU0}_&%
`!b

s})\(yk}9“-:10n/ubJA)$1MZMђwꂖ)/ŋ; Hh7q"s!0bwh#R
%c"Yqcw_݁V36W*7q᡾ް_&\4]I$藔Ħhr|GY$dclWk5;2]z4hgX>F>	*O؟/
mKrX~yƈ^RQl%%bƈ`.[Rƈv}hXx?FՌ{v4Cڟg-火ϳ2݂+ppЮ?+-撼hjXyہ+GP8!W\y1
:v~W$\{Y
+ev`/ؑq;W]+"YNK5Hzxt4HtJY9
:{p~]6K^2TMƼ,wӳ㓫1x	9~hCƑt:\U\~ǝwS|ƍ]yU^:߂ܵՌp We
reqleIܕzS(NM͛*-`}9fJnjR\z7;B!ٚ|55ZKq’rkPLd^Z:cЬcT^mga
Q]MurTs;s|&RuX^Enr3 &Zk/eB
`F7v炅7cv.AFb-`tdP1-I/	0^hg&V_/Hjumph/h#=6Gٜxr*øjGWa">|	1Us=^L1ǒnt3)+Q@;ZFXA}>y>Ү?I1b]a'%+PR4FJRBX(7gM#ij.ls!l+zt/r]fglIZJZ(Zd4-@5#TO!RHapR﹅K.h ULYcO6Xh^,Ԙ-hCrL|
cb
"7gԨ1LuXXe!aޞU]ڎ-@F֙/j%~4z_Xb@[tv,JmBf>to##1>.	L2i_堡@mX3\ՍPtX⊲.qi	И5UKpٷT`0TAp*NlEIiY
ȆvUj#

<
vd[[S#(J3P4[5q\
੖X%CdƹhB•deIc8~'eI>VBnU{JhE`70c:o9!.q@Vch;&I0`LF	J
vƔ@ͼ
`f"\+QST SќƀQ".HqB"-ڲd"[HlΎj@OI!HRWjh,tj<)tEpԳ`-2h~Qڰ(RVP'#9E6HF+5!4*FPq\d%byPĤO2t4.еoHq~1(V8/pLo:k
/m;Jxp4 k
6P `
`93#@JfH-t4ċ6@\~baFs|Ȓ-z=XnV;͠m6t[
:A[v)&YY	jRp ŒCxq9-	4pMY~k~]B Sq1R
۩^8]ōzo(\8o@2&zXICs3V!9}"j'5et@M3!o6ۭ~E
^=G칊MW4{#krAf59W}sZFY@rV8FN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"':<&'ZVU'zN @i	tN 		"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN rHr]	!a	p	hscw!J	tN 	ؑ@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	tN lк&'z@W'?z't:A'rR@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	DN r@"'9	tBNO+k8ފ_'^q)^_'o'zʝJɫͼIZ\׏0hx7&vYjKp%@)h0.}\{-wUf ur5Ķ)7t^4"ɦ6& Lj??dXoN!5pMC=3pRJ^r$ޯ|,M`4UNsRYi]im8BNWR3+cw"2WCWU-W~t(
S+k45hrVBWWRј)ҕ3^
Y]!`o+c+DhmS+o%s5]`eY5tS)SkԂꛡ+uϪWg8G?+돬zJa]tZqfuEt|4x@BBWV6vBc"*t%(<~ףF]!ZNWrk+Еt_eW
;ͫ+k|-th;]I\] ])c	Sptg)TVU^/g]AY'0q{W4͡_(׽Iy;Y̡©?JemX/tQҫnwًQٳd0+h4GÅT^^GK3z}=~uY/hMWͻwc%bZMQWl(q:*&}[~e~QZo_mݾEQ
nzuyuz+o(m7ҏ}fe26*xmNB)5{vI}8eWQyh˲1`aX5tp]5J>zB@&TEte0Q3BΏJ-pkUO2pDi9	ҕ沞fWWCWv
P
F_DO*djƮT4vЕg՛3:+?9
fKWЊ#~(5=]=r
z
z]]ڣmtDW'HWB3lEt?^>$+D)5	~gD~wV?H
QJItut{4dX-\0ULk\1E*ۖn{+wx_aBh6hYV-[t{:}zUs*]xzլAjvNŕqbBMEC

µ@˙NphAk5BWvbDW'HWH#YEt+p-
wfD)	ҕ5ښ`'x5tp,>DiNeZDW諧4-[)dMt[g/Z

9vBշCWUoϘ5^9?a\qd|?HJ3%mDWzKΪ+{lt(DW'HW9XEt5Wj6R1Mtut%=g_u(t'Dd1vBS+t-@3SҞy%F왫6tjh:UƒmN~bc|
7j|=IY<-:]|:2w?R*YЖtD(GÔ:cxޟ៛;?c}V^sc	|k7';?.r^m퇛<avGɇNnx>{)]|=>׮wI"l͡eXFN?/X7>|wW1~L,w5]CWoocvqbz~ĥgYXt\])#]`,wp/θqg\3wڜ_Wf{h뷌v6GNܮ+Jgu^0Pβ1p90("_^}f4pdbufCj xh6c5lvxEcʍKc"(LL5)<1U~
PቩfcQxr8ԁASkZŜx"	N@Y,g]L9Mk6*x.b8JD	ptI	)*.BlT]P'w{!,:eyFإpn=4xΗu2-}+8PYf"'K	q]X93߶)2b5fZ逫zd놃r ^_y_Mw0sBc]]|__j1kHaGm˴nJt)~v𤫫V-(Jx.KlѮ~u4E
ZPʷjȷThD]ʷ
e"Q8bN98)gE:ǥ/ؠrTF	m-,dt>\pQJ]L.B{Q/m,`po˴%HE7Y@usɗ/7ovu{uVjHmn&e1іt7:dv$eP ::Ǔ!KB6[烗1ZxpK\:3pNpTRgSg1d43<0Jl5
ebQ3MֆR%kz,OY%- yyz+l¿N3ѾtBZ1jȰrʁxP	f3L1L
N&VRTKo6\Hl
@6WG9}nx40wjGl	a%AS{Pqpds&<|q;w'M^/&.ltM[*|Ma6э#|y-(aqQp3yRX)zeJYd# aBt:$g>66({I,wd@ϓe>Yyҩë%w<~Un^_ob-Q;C݊OCAu0?3s+Ά{+v
jOm*ꀈ#n&':?ξ0=ZGg&Q߆!WV
_.RY>Ht}$ٲ2ѵlIiIi9cʳq7oz9XkKi͍%4.ԤVɩ'rˑ;ŸAc@,;WeS̫vRژxuo㖷޵q#E? Cwrwnl2HVV"KI	_%hz8 ܢ>~Ez_i:ݍHlQ^e dNcJ|5CRH]PƒbБiL=_'ȠON6rpySB䁧(
rMAVWayM<8O._qrȓN-p3k+\ͦgi^MP
NNgu༭~S^]Fp:
`IHhU&F{~|_PSзw|&5Lb|,Ń(Xf3&sX󃶬P>
|]0QzGb|	o|QpDk֙%
+[0LWbWRBqTd(7h&Y0|bBI1ߺ|0-f7HmyĨ6ay`x>}߮oVF;]Ӿ30~^͆2gsz[qewF|7׌#]~[EmTs[?y6Z&b
nxt{ut8dz';'tlO[jxfXoy=m>L'-[}=yn`}'^Lt4nVQ-ת}?ߌs.E#sum}_i{yIërjVq=Z&h!ܩRBH+n~,OxxGi
Ј\̮I)=])}yg)-*O^AFjU/پ4d	s6,Id\ĔK^*DT|e("Р!4d(EA"g8Hd+;(fZIעK[iFG
)Ho)c).DX
)@N+cgF} 0zYRl44+PWEVpDPˬd;M*y!&W*%a,%N:-)iXd<sN$R Ut%,"pkFt]veO{&?KO>^=f=BcE\P8j#*0k/.6|A#nXBl㍖.:bO펔)fN;SLSV'Ide/FX
KBՕ(5&LRI[J&+a$b۞[Om:l	r~Jٚ(1HXE
Z4AEAlHVKqdmJF@P"$[@	l?[Y?!HLb#
.YL
3qv*͔i)M&HB:ן^gnBsuL#z'ʻ7RYs*fѱYC3r-)$/(Y$|8$7GXtmn(%$Зb":tҚXF#3vbMG[iʌi垙<6?ŀ&F_FGl0u.0_g6t,eGKZvn5#Ï:{))*lja1*-|(2LE#vgF4㚋y*ݙv
j{W{U~eU|mg(yd!@-2w:9y+
{EIʚIdH82F񁘰XL<]p~2v&n٣au=݆Dߊk#I!:ƛcͦ$K=t<t6h_N"0(_|8:jZmÕ1*eWCMʻ?,f
1@W|5b,^(#ruO?Zk|`UceKEP2ew+^җy7:2_Wk
sW敔bug|/|u/^}l0_Ns7Jc=;_߿?
ig>d$ss֪/~]}33+@-CWY߽ZC>lRC!̑mW}n)ݷdtKsz{91s;{wn-lb-:9[x_ѻy둍1wЯWjc*k9R~㊱'c;a%~ͤJ5>u?/t/y?*M'㡦qNdeVwG"o2j<ώ,S-fƔ͑EX;wq#J.wv7_;Ncp?]
EbuGSA!zR'YՁK-g-^Π$CI£c"BԖJ$0"GۺdʽӌӺsJs^nix
77n،|~~
":5d3(PĤZ*6i6}owbwdև|?4\lhdm<ㅱsfztlXd&iDcjF6\"xlg}M8岰wߑG\lv=zgpNmt֖/(YԶs6zD#:ET[6^R \ʚ]D #@mɇlRdq䐔
 KAЙ8{\ɰk\9UTI)lr5ʹad6/q{vOx
52rkV_s}}\*VfKu@ĎF2(c"VT'@HuhOF]%y*拯:hVwUug0+p.ـW`28⃱'b)Yr<9mq*+kWO:⹍uB|^XdOR{ CѡN2$4O3P
%(,.DoKϢOMkj+k
l*&:І]$AG/+,$QJAvaB`C7n
^N1I*dv#	$޹qc{_,7#Mbw{M6H'V#?ji{aٱ-{HSԉ!sΐ8B=-:`HH#xl&Whݫ04{ћDZ:?u!q2K\$qdgh?l>5?Jb~MTq~c>sGbL
8!b[6>%ZqNy<|<ǙV6
OjyRDx\-ȭ?uwmH}!/<T\:#W6:u!~"mXB`J0K?c
qQp4JR+.z,z!R>	vtMG1!,yݾpm1RQs>Mr{qm#en/am.,:;τ99;ѮӰ[i˛I/OZ>īB/9=.nhfYF:뮶sk='0||rqtҌtLG'KM_`L2LoG^+wye}o`\E0)wKʍ˘
ۖq>}s\;|y
EcgLq&@.+v^{_kh}_//~SS֧Fb^]ۜW;5R_$-B{ /
u{s՛?lMthirpq*ojfNz4,k
'`7^*L{c`.U{bx-W6J`8
cMmtmpSY4
Q<3mLi8}>I|ge;joMfm/g99/Hbuʛ/弄&`	'g5@bWO5ڟ/-,!/anO᥿ǣlx/ax}aծ{O%Zp_21.؉/C&K1/yFh_G|t]wl>v]^ey^-m,߲aUӮK_'w~OmxN0zw
y7c7Y#'盂X42?=";svٱٓ)lI]8)f"vUO!M0`fg4,g	^y8OR7>GAܼaX^>Ե!P7sј|=GZOo64®B~r1<8x9.]fAִM\5&kdj]M<5Z"5qڈRxURuE.Ri'7&0DGfLTM nw|zɢٵ_j^WopOz7ųɳLZwJΘ:9ZZ$BY?TD9ᢨTmJH+G^h]|-M۷d266	ߒh#-j&juĢV[D*E
=/p)ez%3w-O?L[rBLW&k?bs/c2>zBM$H8U^AC<`g~PX]+.
+,+&ϲ
޺\sZ˭hdE
j)TE
TJ"n2p1W X?}
2PVѢDq*) ,Dpg:5l@$o]J-ތ	ъnWꞭnj-WTjL\v=D0P&કksZ(@,$،pW(W\pjNWSp5@\q,Spd6B2\Z<@%XWĕ \|1-7>ڟ|,Z	pp9Z*'qg*Ehdi۬˗5B4AH[?hX=uOior/U*4IBQ-Ngu/xէڲpHn1G3Fz5jЧO_~OK;YToG䭦j|2z]H>O	磳h4>#+
#'ڱTR%.>{Vb€jBEFX48hazz
۬Zصe_w߾x&(tT4V^z[5L2:gemmFp/5UԼ43WkuTTM}t{I#'ueU09-Nޫ5祓/Y?^	mRq}%UuuS6ԁ	ai0Q`Aݳ	,\Kr	,ZQP%%0d\pM>mP&
j5K)3jRLIӢ>kx.q*u5H\if
3aU
:PҝpW(8#g
P-K޺Bj\`3r-WVq*yYrtpeovLN&WdZn*Mb+W]O辸:\Z%SL-Qգ	W(wg\ujmȂW(lpr{wSKMBXWCnJJ9>b|"Ke)gZAסB+&24
VX(:QĴW f*:P!Wĕ'3`Cʵ6\Z[+TKnRuu>B6gRRdWCĕQ`N]\nr<'Rqe:' X	
PDF+P.z*v]Oɘ(MU7F~qQ귚TGUڴAJnWwz9
]M\s.|(Peb
WLSxFBg+D.q*Wגj@0S6\\s4y
UrQp5@\	īI԰WW&+er,xi
rmefS\[cwR=*\&xaVeѺlEKn*QЬ"L(ߵKaJdZlB(W\B6̨rkI	-' 
d+,W(\pjKWRHq`j2"
lVZu*UqP
l+P	IWECĕD
Pg1U$:u\ʲxV0@0

͢cT]JS?\[v=)q7G2w
=-:RWwzj-2#\Qe6B+T+tB\
W0jXFlpr$u\ʭ-W _`U4\ZbiЪ'TNsa<tbn6,uL%
iHFF`(DRJSdjRF3U	9ǪD+TkEjقJ[/rz"
eF1;ku.Tj[!XnLN糀
$o]W*Mrz"j&:e2g
R%pnlL	y_U7f\uk{U'rkQLbp
6u{rm.B]SJW 

Pf
2BSdjy)ffl*[)YmJMaNCVlX,C6݆L9Nti|)&4 T4L*ureJ5ÖD1>k3Sӡݔb3
-`σ+u.TUTZTq3V4H(4H(5^G͛	]#@Q
B}˲>Ls6jw6K4gʯtZ:1=n3s3$@ǪNiQI:&0	s&Mq#f$[8zhPQw5qH0t)		$ᘈ=/31Ok [i"e8c$eet٬~/_PQq囯]s
$ݻo^gWH/GnYr9{\*giV9xg{L.p
S155L䃽3\_טm-)
gzЅTstケXjug;񌦇ҾO1vu6yH(_/
<]hsts),[+
\baMq2"&j1h-[.q\q$o9W
%ʅ*C
]_@2hCWGm2c/m̭i\x(.t-c6ώΎ#0Σ`HJl@6}j`P̡m"|ٻѡ*W,c`q]bNqM7J*n@6'084		e>ɔDT?BiZU`Rwo&X^p[L%  [Wѻ;d䇹=*uFnTzCݕF3x@LAA؏[@(iMB‚(#2} <}QEOM[Qj#(XYHXu0wLƃckY*)ԙ=Ă*qnL.O
/-H2:@=5ꍁD]`#)#y՞eAQ
2"}#_ ()W:BE*ѤCTݫ.Rh=9VtVd0WGȷF;˕K^Np5dYgCFTRG4lM#FP>6i1b
"hq!:3' ]
^KMAIX@Q󈦉t"B!NȘǀy;3\"˚Ky}&;}-x
acGf;]`A̔ZI|t(#KyD*}t*Y\#$]+Tu*3d`Qdq>U@J$jC62q@Hr3P&Dt_5Qq.Aݺ& q<11Jrj@#[3PdGFby3Fv1Ab;~b0Hn9^lżKȓmc,M&XU&oT&YwPKxHi:EjC$*[PKbib#nAQA~,ZZ5K
qr,-ChW׊1;y 
N1!R-
{,hKٯ}A:G
n"h)r&HȹӐ Y5 čx,EJ30fd$?=k[1pQi|l)j̖^j#riZmȎDЬG5|IrPb:HP!U~A+{
PsHA~|AAxK6Bw1wn7-R-5۲x6d?͸z}[liKL@1L?٭_䯲78ZR\o;ٵfv+粣xrݤf$/>.gwo"[.'J?N_ BgV{̕\i~ur?˳7[#~	myջ?^O{mNk+I׿P՛ٹ3a?*bN] E+wqR'*>ͳ%oCۯvk>[̳cx!
N
gj֮^6~\pcٶ5oŋ[e|>]wV~vzsmʫDӖ^8/o{т?˷'9CG_h̟N~<{nW7HoV+҃\g~`A<n&/׼L+8҃F9(ߟ^n9ՊQvag祥dYN{ 1tAx;ⴌy9W.9,)efFd:_0Eos(-6SYtTjeqYyB7(D;~1ܔ_pa==`Bua\SN;?'?\s;N9m
!|/owweХ̴7sYFؗeb]ZxKrҍsF2JPRzt
q2+RW7l>
(7OA|JN>P{+t%h#?vzIU69-+Nn;b.oebCW3?ǫRǯmwmJ߃՗=7DWvJZ
]	Z6wJWO,ӆJdfJfJJPUWOpǽǧHW6S
گeJVu
g{g8sxWƷ;هˇ}ҸGko2A ,lsY	߷غږdnR,lvY$i0+K[ܟAjˠR5kqB@:g+mH*wr1=%	t.I\sIҪ @ع<΅> J?xPV⒇|;iwJRNG.p%(&F%チ$.9=IPk#\=BvH
{]WI\~0# -b*II!\)> 8y$$WIڇZ2'F#\iF=$JÉqwʸ\%i޳$1J+aӋb\J`mI[fg n.&߸1z;U&@7S.g}m>w:nꞯ򶫺C;6XM$zYm*iYu43ԕN'7߬m|=T:Z",ж^CyXasz7C36ӤS[f3Tڴu)[]@3RNlg,IAXJwabVZ6ͳȷQ7&y5y4`'JJ({4\}/=Qvꉪ4ꇵ,Vʾgw>=~6ɢ!
^p"E޴ٽRw?g1&}F?NsG圑CZ^4+.$i/^ORcc]\yH{mꃁ$w;O$%Gzp
L;_ˡm΀-7Hj	KOtyaM"[unn>/GI]9-6yLdYs#ωWڈ59}/YtBI;V3UVWIkbuazA&u};6:oO~gD=P :T`D5(xO8\R—H(UqW_xمMw𫒔z8­@77Qgeh+esC-L?ݩ"vPmqJj)ٲ@Pʉ_\fS*L1Q#IXG P	@lP)"EKL9,#s`&2p,\/`
C583,bzdcN4^JJqJ7Dd%0l.L,Nj.:Nl& `"ЅПҟok*653p	jdw̿V:voL^p~K;+ꪪU.{NpDی׉p/*fh0G%ReDJsN(>mIiej_Mr&0lZt㪖[H2-m
oM_tC9ikPK3māJyωseE+{Upa"ᖖ&?_0A57Иvm/z+;_Ʉd۫0.B	˒X-,H3D*'--[
C%Su7b^2GGYu{A19%.ól-:+CEwIMG_Ti͕ ZGK!zƌƁZ`1'+j;+*T=t͔—6TxtGv}{*'\
)QކS^ggaF2/Oi,E:d߆J.}=g[6mu?un+,m	_\qKd9B
9O|C^*\U`rk(ԔnQ;wIOo#ko=nertykPQ5oRmUnѮoG TSQp.7YK5/7~L-L'wDtPJS/]q9X051eUMoEjΛ} A;ȷ4NP?1ĽGP0+NFcRG(( n)9LG
k5f,`ZFDhFs+\ÝGkga94t+dDuw}wai3	KID40}jջIhǀU&`QaƃcjԚFfA>pʰQk5ia_@a#T)JPkgR[Zƙ{Wwϕ25-sNM/8ҿ^<8_*(IzV{A'e@7<[Sr6TWO,>wQy$'(5G`Ita#6Q`;q Dz)P2zEi#)HXp).}Ԗx93@ndMXW
kӌ]׌숅3RنČWpÀoovp5p`u).>?8b+)ZxO._1\95T "VD)@JZ*(3VĦ@TCr'ZK(IItPhHHLBD s͈]8]_q1kӎ}Am~D#MyRh6+f*&DeELyUF
sS䁶`W+F.qVqCf,
G9AQNq=	wYS?9K/X~슈fDGD<"moӂkl"Ƽ4Ef@d
P\Qxà:>lJ֊Zad(QHLd$&:łKsI̥6q6\>pqOk:kӒ]qQqߕ):X?F(A!X
Iĵe\Lk2jq>pqoxb]ڱ#V#>jCV/kp5\;xཱི	c`nco#D/n>E^jF+!aeApLJ~;lU; ćXC`'Ioc`GbNko]c7ٮ8cw^cc[/[k~SlAA1LE+b0+D$jˆ2&֣,J텰;_2fT9FR<be}0z)#"b1h#2&"] Gml޷ƥsɿ\>mac0>:m{pO&8hUW9ޗ&=aOoy$	!p!32JPq@Y"
p`%VpŽp/tyG`8Q%
h{iQFBy*bVwv;+bibLޝ7մI/db_l&d|-HrVx!vXQOUCwƙBeEV@(pVGxg%GCAHrO‡=>ПޞNd(_`OBAEXg>D3[SLuEtd0L?'k8-rQ``
oOSհkM}jiOrr9ʠUW4|5'h'pS,R'35—gj{8_!eqo-8dnsw/cRD*(Ydj(	K_ꪧ1? ,"˛Lmۏsӂ{,mO/VC;!4g'5fSՈhD(jD(@z-1*pyWW=gѿ[xTՆWdӖsVؾX$6G.
ދZ@w쏆zK?b2Mܧ:-3eu.,LP-=Zkwtm2H|ۺ]ܸ.-GM&hqL|?]<|yr5qO喆&vuwܵ]xʮRsXr}*95cm(tphSٓHJ\4IJW@gl̓'g;Ett|WxMM@I0Nu"kBV"AFaPz(SyLivqN-.콮3ӂδpPgODPk:}ZeBt?:Y.FGc%υ1"&B,9YVv_D*,e^_(n$[}̟<]ֽei'HoG+xmw._fo:Ũ
dչx$fvC(%RѧrvM/瞯;ba({omv0aȄx,ҥEDsv&Ԡy5nw]6@XMI7A؁i_Fy\M{T,X43yR>9+#XF2۠L*(jo	`62А샆Oq>Oc
Y}֔!< l)EBʎM2t[6g-D$%TZ2L9ZD/P]&_QPGm
v>jnخXY?~[J<͜vj&ciDc*W]!kaHe?+N^[q;oV:}gۚ؍G!h@HDIVw9@A UZ"rwhc[vacDDy⌎'ud3,lɒ];4,Wڴ7PtWv+<'cO3<ц廫ln1;T}
Ԩ
o:G̚= M,!eA%AH.xРu,-Ì"X=E;2N;bn|m+cƶL\Ub#mRAd)4	FLq@XJdY
Uػ^tS]Pp҄e!HŢ@<ދ	J0l(D6(b^6
L&DA,$
[cmMJdrZ#	aIkQbaFl1X'AVvwVT@[cс	]6Č
	Q
rs\uĐUkYHGlXYN*E0!OopFW²JmpX9W0&hRBA	?9<~ha^؂JZvTϴV{2ұ!P2BZ\h:Ǔ˃:>j5kΑt24}o5'4ynȍ-U.Zz\[7Y_,LN[3Xg⧓DO϶ت[]rS7UާԆ>bJ2jP}PPkU(ώy9ћ^|o뇟_Woro^׋7^~;_xfxS84.UZ+cV֕Vo&4cenT늮+uƦʸU&y_WOGi<1tࣦoډMkgӴrԋ	O7v*IsBz] +=/AxnMkz}tM|$ծr
,tZKV/!3Ob(ELM!-[GFߙz럣lh%4:RfBb1H.ZCN&J*[T3Ntzٚ)_ey;zPՁP5og8Ș\cl60<܈TڥNN!G4(1Ĉ?!iѣ5!e@`I+/DN0Q\'SD2TOBYxFVs28b8?GvD
׉
9ubݝ=#7>ޗB7z.q@GOvfuLpˆ؜Lซf`rN7&U#
{!cC"x4HbF]B DUZ11K>Q.dUIYzٔB"
 UNNʢĪ}1S1H2Jc#
-
]NQ;wCZ(3qH[`}=GOnDЇ
ΡOLRN89@P$]cؾnc-NiSxylLJB.SH^s_cO1
C:-fA3rdЀ:TRƔ j89~~mĠ%[X"adeᅢ%F^1:+zMqݖ*GvYl{~r.f{E_gg. ۑ	P
fJ!km
%*F@zaXu]LwK(-y)A&8mI?DrDT
Ձ$:κsߊ9>?=M	k|
6m֢bWYVڵZPZodCj^"uĨ6E6z>6皫<#2L_]1~0wSw?^ú#O?|z.k<7f,mϬܳ(2aYrn׺25Mg]UA[?Kb)&[r'YOjZdyk+7
mw5яo?oRF-trvΕ70z*wL\Un?<j]^ʸqeX |^N=#0~>J9ѫ:SK>kb`q!*@9Pi{#ڱ/_&fkLIfr,$d|v٥,D
eKXڇ۪թ_kl;-pnTK7k7ƹay82^6K gKaDH@Y#\؉*ﹳ5=ƩDP#z":->hg3:Ǻ~=tfe!tlk&u0!X6@a{
3=3]Mfl_aa>h%=h!nY$J$\(hh-BфN/A,KdOґg`Z*92,=(A1\;H>K0֊(5xn`ۃizWJv7>Qv2:8꼣us31^}"U{*uEKx	fF/BhLN1&\B)9Ag><#'s
9cX4dI1'}A( XITQQr$HjD>dkx%Z%tFRVฮs7\e?FHWmYtAG-6@T#U‘t+1wV"%,Q|> GR5Kk0gZcesPx)Z-v	eDj",]߂PoJ;6w6WBQEv?5֜,!(Yb:`~^,WC~XVּR,0e3v&(ؒGbi[-vۖNYdby
z|ޔʕu358nsݻ{ӫxQdM} gxj[z((vonLD|͍L[N5P6{Q;g֖[|[R["Dǝ2he\fwr>J0|NQ(e:e਍iUCYډOt]񑁮ŶCZ2Q^ieYUV&Ul&r)HYr+Lz9ӎ
 dtC,gg6KFNO:.aAb
dC>B@\d3˜;	iD'zKS;>Ee"J6hZXNJqqx9p(<ʚ
鯌Wz qtyFyR'~q\,P%]/SHopZ'.^]~4meREVѫRsd}ANA/F׷|$ۿ{		͇el[_,dB`.Cζ]#5j5۳',wIe>nD5ooP אdcQR~sV/IN꼸@²{iN#(w5t0ຳM.wrwfyPPƨRaȨt|S7DMc_G;]l|F
\{T7
	n&S(vewRl zZ=^S!?[ȭMJJ:Dŭ^u%uO'k率tEլi8.Niu4
Nuj
zWV)Z·\Ud[WwTzxg牘l?1?OV!9+H(rYxuU=pIFH۽čRǔO%	}IG 	ܐ(
S$9>5qv+uYd^l^˭-:bH=@A4dQ{Z@!3NE\-w29a9*cQdCW"-ą@9ه`QK(`p
?gg)feg3ˀe4Am$$Zfl
ߚ)E%ڢ_9ûޙ&z7ǩS6A)c'N`>p58O
$o/{>N%Gl0L\FKe<	^ɔ	jlͪf1J韉X,6ZtC6a~F>ayT.}mDkfӼ
*%YN^E	M+Z'h|[}͊5ޢ7Wӳoׁa@s>dzJ۵͑p4jZ'z}Ko98F1Ɩln颩،YwY,HiG}*	F@OVm.-36*VljpZ:)Б0hKWr>?lW
*4'~U6}W~ÿ|}oh3cI/1K[XL~3L&3/V]vW\AY	
S/-B=Xy<0X6`Ӗ4hZZOӪ^|vM]nhow۵`V$^ڄX[ڑf|Pd\{wN*IkjOJ'!'Lp'BC90X,@;mQI	қ
R!n8k
!#\ɀ&9#A!5 ;#6e]wʙWڜ84~GPnwMKxF񨐣gxIYSڿTb2	GApF+ZPׂ0Q
XFXc8٘^Y8A:@v-iI_v~CFZd_:{>_Bw-:3YءWBGϏ2}&?q!D·pFhӕVVʳPD2Mku)^
KC"x1Hs	B22{$h}SɎ(}
xq94VEZ84VEJ4V_c+k3PI_%༜U\6褍hXdRVJGW9̱
 xaut҇q
ג}U9x\sހPT;pk*<n:,tuB*_-{,/9ų.'Ҵ
G:10}y-1H3;YbsD!UTr2(~Z&J֟;xӺ4ʺP>;N'Ag;nUtx\!Z%D^>o@%*BJ6e%ewNڞ<~@ w
w0OŸBT)\d (	Xq,?,0Faŝe,2#MID|8Ƀr<&rB%pfbhٍ;P/Gxs>~{(ˮm2"<60#1GY,i`4|_9=}ihv1ڔI	Ji^pȴdxZge`Q3&>9z7#2f%{t
S(mX	.&!qッ2^֫p0u3FYIk1PS9pU":dn(̩]NH8D9PeWg
hkyvy3|@Xϵ8;MZZ;.1T\ȝ,%J61Sy@gRwۀM&pc%)k"m,fyRIx	eK$3cI+JK@v7x-}JiP[=E9H{Dgzf{K6h0㱶TSH&ZJ'qFz	J$$AdM/eK|n#!5\D)
nu[
{&}'H5q./ѪiTǤI\4\yYF4d9d9=7
VB$dIuN<E]/dg<%ܓ]W͗zhU~oGԦ ll/n{٘{9p?(dSI5M0q:J\p!EPNu@MPz"+P%Z3+a`ܖ{Z Ϥ-6j}hZjzK6FY{YsUvGWY_ru2smJ{םP.:OCU춆.!V˓Vkj
ȪW?z7ﺽP]ȋzv9%OM܅Uy'&-w*^oY‡Nf3rPs6Em(5.T]6X
 bIeQQcAV
'+(󜫧<9OqtVRM?"]a1V+
wI<'VchlH#<P
i>H㝊4t9dj(eC)7Rj0[Na/
|(Z)"B+ٞrˈs9{忍~_|&[Qϟ>ڦZsEH;CP
JGF	7ļ@1!0?	Yb$FZ3۟w"A}6XE|TGۍG/GS}Ʊ92m*O|c7̎!8	;e9@3
1@,ǔHH*|vU4	\@ൡ?R1.Ebh18o|#e`T<j[#T;\8[\M 9Ơ|DXv"1	)jU\hr9T#nJV݁LA^zm"{t󺲼z}duTV=/Wc8Ux
+|9!D
!.THE
|UP#pBtf7/^RݏUe+Ek<,:A3PNe.E>]- ޡS@K&fQpb(!]\wk=PEI1)*!k<yogPf>7beOd6NڙzF@hvFkpU{p-ާvE彦QwQLS\Sz}]*ԡisU`w]<
z"rg\R%WMB֖+t|nk!׊V|U|Dѷyl]in6ӓ**K9hl]`u1j8rMm63Y;
'ںm
ҝgpk!W\}7>oE`}tJMmDMUsImIC[CUqn@r4\ifᥜy|հ+4hy8)+73ٶӆAwuYճ7<7f{œ%AY.Ю
:hwݡ
>P1@C1C 0x$82Q1xkSY͛M91qCӿڱ{RuoB햌jg
ηym%{Y6g2rJnΕFy}KT'K:.*in&o-wގiSI{SRgw|
V5qX삧oȾQRpFNX6&QmH).;yY#R-,-"T'p™"1Tԍ],+ӥS13:z.|BI<XC#w&@8SBNx`:MX@=aDL(=z],\s	]|
1bD"mm7TBJwKO&QSIw=te)4~tGh>6v*׌?.6<(_d})(t2('Cq3rlu!e(\߽F䁈`dU,qAʸ$Ḉ<#@C~β+ u}h/1蜎>sd
NH5`QkPA 
D|&p’SQZ'j#F:*q$H߂>I7@)Ey&/UЊǐ56=)k'Yqa>pQyO*Yu8řW.S'9'!L`4ȘvRU
!9G:4B(QBd6X[J؄ U72&k*aam+,<(.Q
33^3:eGNzG
>t;Gl8'+Vhv6I@"4$\2M9.9VH!{60ceRy
3DD02\3b&fvn31kӎcAmѠv`iꛬT¡&bSL[$er4^uO('iaRYﴠ22CdQ4zbMK#b.$,mpkg3VFE6 bm+"Q5 o90(A
B<&KМyaFC6FjED)!RQ&IEyPReTGKVQfDM-,"jXNr^gmZ+.\L;,I"k$BHu=("lDPU(ФFţuiǎx8QVcb]%Y8k?b/vm2MtzI:=;d;To9aclyRcP'H9An(DkHYAae:*\ 	4		R:_Ƨ97=t
1q]6!]7L,h5T"=&bjjgK'r?}T-JǔbG%c8i>t}WAۗ(ǼWMOeA#@*FI	)'ˆՒ˘
]/?m:gbz^x~|R&vQ^*KA[CysˑO5!w۝߱yxr@lJvh=})-P'cx;׃8;8|hKw˜81ȩht\b"ܓ{V|:Ry6FxP49q"7;{ɟ[n7k㉜vC]=zzOxLA!/2fxA8r{99?#*]WQr+C_RBWU]2bsā43w_迗zhLes/(Wsމ)&g	fS
ļgm*ʜ=9VV]7wtem<-znq4#0΁;_	ΊEd2b+̓7Gb"DCEu"Hj<Ȼ^]]zvTh#adcS4FXp\ouCFW#נola0uuٻ6$Q"`avcFwu5M$%1_IQ%jQ`23ϚG<Rݷ\.7zvԶjڧוA)T\+[;uvn:{Λ%
Xu$#TRyY\ض6WRY>%7%8:୷AJtLB9C] ~st,v]u3qn=`>2qe'b#H2"hib{{[\tJ.VM//},s̖&V`DG
Pi@3qVB}SQZxoCM,90@ZH>9'%6Z
MvE3qj _&!KѡVLsPd@T QH\rd[z|&BM^1P'm4d%#ِņÊ(d0((YLC[,aY WV*r:жLdHR;ñ޷oQv[.uD
3-iK.CsGxBC<@zZrJi bv2,ӹh,@X9BkW	8f!Ež
eѥJ]!1|Owʠ̪ثrsD`z4xԵ>ҪgY=ʲnVt]"]q<=2ҵ9L끘&՚i--B:)X@:+.Ѧ)MD'<|քC^>+?R#a:r?aJ65At'{I˅ܯ+:p?)7tX2E-f$+wi
	/k;y˯5x6lճKrBjZB-^㞗s捧!.z_z"W-@ˆx~u:Laa>XnI7?OSYwDhQڹ0&
!*ژL$Rvᤝ$"<B'b0)2Y`*1 rAR"!
;]:C]2#'8FrLZ8pfA+)
Ygl;ZJ:Gx댝?lR-$TrV%lĦBhI2p%u&}^%RRRFƬ&FМ4dxZg0LZ4<e̴-BdnZAhC]ZiULcb^RfftM6P+5D>ȍNm5ÒrgK<yƕ+i.C)S$dW^琑)TX[2FM%(aWx^er5lݰ/>r$7{Ȣ=*jSՄ'VV\uVUN@.5\2W\/E0yO*;b=Ĭn1dTj9P*DLGI!&霹,ǸV9kd}NV79YW5	ci\t}~н>7 1CI[f
-f=#؞s,|lסA
+	⃯hJbWV
 v	v{JN:&'m"-W
KX"@y.cRq
R(ا^=`>~b?KƏ'j>fKf
-0ZIAB7Ay1gT]|0s:pM&ߊ_@9&d$pCZ\1bdžڙ8
5TjF4.gBwE3xZzL 52v&-B:*aag+z,+^[fyԣ9.ox<4O1!:c3f+^Y4fFXh\
 m۠eIh`2!EM
lhB"N2)3cLb
jwڪGޤoڨ@N~"3뉔e.yq"aVxVq !(P&CXC-8FbVɲ\{aglک/a|QL;ӏ]tGo.ń!Z9l%`,{U!#9u|ؔi'
рgBh!餹"'-Ho8`	p;%zՁpqYqHkiɮh{\qwe&FfCl-0>xd,!\emDFn"A&j\<ؕv숇y{{=OMAp}E?Ơnk5R=[@AvZ\
Ap9JDCܱ?fNG~ΓAid
Y0+Z&Nėd/u6.jv}d,ASHYrb躓 tiA.؉42}fc3qn9#.^̦ԓu](椏'q7✼9x6y՞6TfzzDzN>[G`LJtD91Ւt\2YIHV"oJ@{AK5iLHajO"pMvLɅ6E
r#
[`2AN`RfϕE]4FB+/7$
Rdܣ:g;0LB9j/a]^ٜD-ly
.͛Dt5^j>}7=ܪc*K9Y;eho|5
hoX:vƁvl˾xynkժ+8R6lVJ/R28qvԶ8w4ˎw}0g]*o^=ilo[|o׌ɅkVg>skRIxDHO'WwmH8`{\fv3ݹfo_ƶ_[eGe:IHbfTI#AYtiO.҅mjf7Bkj.ϖGvML2gO 7Ħ@#"*gq9mT9NX&(*?*AJɈ@:.ZeJ۟:x8GǏ͟
YR/NG3';=xQ6:'=ys ltڲEk\R.WKvt7ʴm_ |DzQoJ9~ʈ&9}SzWJ˽Uj%G
I7Jd|=ݳL:fJ=0".-۩w쯥	py5uq~:ܳHfk:hVjn-^FevwƢ4%=7T4*l_Sc-BpRN]𳞛>ZǛ7	!)XZpL,9duq"F_*^d%2q(dx>d^`ڋ^Pɾs[J{u?K.Ͼ_wow2 ַX޿~wM/uw5qr6#*9aLى tdr0%}Ѥࢷ<^I>G0[m	s ƕqfvtn/vWG`JKlfp/h}~yĊ77%tL+ǧlV{Zl?7W4q
^\9]]Jzj4<>:qytff`CC
6
l1	!*.,;QvIyȟܲ\`d;f2FYG.f!|/5fX@mYQOrxi;N*}#9oQB9O-|8]?_P49|2Rgfc4g=
z:vXUZӺ֜ /ypZ]YbcVswiS\z|cYOgCuikTɠ`
s|d`G{j+Y4eYk`9-Rj2@s,GVHGOpζn2%#$+D·mP%{Dj,IqJ
)I(.LHxNêB-N(oɹW$džB㇓f_Xx|\%{Ͱ5ohYr*IͣF\@$CJ&-"}@띖ZʐEJ@jVL/uE3'7mϯ/=@Xzڈ)i#$E#uYV/{zl'To}$'znEl{!`M!n&=@N"6WH8ouA?lMϵ5mu7J	T7 Hȝ1&X"djWva!]H\'FQkiJzdV	zcb
!8H@C"3N"B	_Y\3b/HJd'~UBNW~Km2=^Ju1]OWpRNK5'胟$S8m1v9p3>K3ގ:fϴN#n(&8㒉Kv\#?y8x4&x~g=o

Yݬd
4DWk#8k]-J!f)&ܼ]_.|4CީS8iKriV__/QqȎ%QܪU9'MHsv$'>Wc;muŋk>.uO^0ūe`Bm\s=]Jw۷KΑb1=N޿3TI5U5Fj',6V?]t3e}wzC۵
y#պpZ:-~mRNjX^}3Eq̏s|kO_7Мeγ< uW߽w}w\wջ7iS8Z."V8.ta2)迿XN5>ίUg54)QZP؄@q!^cܓ;ow
 'U_#Q57ZZMժfgMPkz[USBx[ =/@xj_5>&e#~6ZiILI8.D?z|"&օ1KF?mWa\vGZ+R05Y0Rȕ8yiDrq9d;*p".33*g:6'v'֤XoXbq6^K
X5&lGnޮ%lO]!GרTnQl6(Fy0K8ܽ!H?JhCBi.d0hR8Ra
LFo$b9TI9dL[5(A?\2&ٚDs 
qR5q=WH_H]j5Ķ'|fewfd;*v/y1Ùn:::t4.g3Bl?mUtFyhXibA=D!~6HsI2{M9O܁TUDRk.JtrcbYPTˊQJG"Yk	W&s^yB}Dz7HI>#}={BA7-d}&.W9;kuYY"qcqA+Ng$Y=Vu(Y@x\%aN{2hǕb3-c8(

$
_sulش)Qύ~-LG/6BҴ|mbb0W2`ܹ5Hф":hQ9i@V%9~~L0&/9YBū(ye`9!TũA+9Oڹ7glmwU=Iuݕ*d}t~<}>8hhɎ:s'	L`ܒ9)Z21VU]q
t<(,	Fa3z FzZT92_yWg}:xٮzܔ`MZ(I$0cm$geHQfN(r>1aV[smVjsT^gyKYeI	TX5s%Qhhu9ig8폑.cg6qȄa%$11b6[Dh=ݒy'+jlG(F5a2q2,A{(1
D1&kZ8A*|W؏@HZ|RN(ʸ52Ycɇ-V_=Y6Ղ;VoCm)9~ؐʜ[צ:leW6+"+˙;TP[	tے3
6q%ZN)xm0v"E2RH%xbJ')y`NE}*U>8nE6A	UNBz^kfE48LPrWn]V]ǸC'Z*+gEn¦M,.L`%Oą"_?PGYӀx{~Fؔ!.:\&O0n<	`0cZ@r:K(4ͦRMxBW\NE\ejAu]\e*%+WoVWH-qz**S+tUR^\JqeJᓛU^vVKՎŭO/+POΛj!A,'~
x#5eVp&:%Qs㺯7q|us͊?HseС⿏pcSx(}t\z˝!z#JIb^pQidcƨoFcθO*VT_}Lwc;htg?vc9J2)jOlEkE.揿ٵKc.8`@>DQC>DQC{C>DQaǰcQ	:^}n`vA;r-8^bEEݰ:FpQv#-lQ_kY˵FJmZ
1AxGUtkDM*FZg[q%!MdE./*>]a1vl'B(>\IA|ߴya84lenvkҠݭ`6~1oWŰv1f.f\"ZiukUkj%oWy@_=)mhCo7-Ǽ˞G[AڥX=:̕+ʮP$w5ß7ݵfeoy9U6m\n-z3nv26E^IזQ@YYwxyKOL >s6j%/K,KN}j-渎kCk}Zo	JYCE"".x#3XJ
S4eSEnחΓdJo^{Mhkԛ_ȀrzB3t"!SI=n8<=bV1=b)qۋIm&7}ҭP]I%Kx,Gg8mw(ܞ֢)r{FYwCn Ԇ\XHԔpmIomP₊|
IRKT4"C\ {ń6KP%$	&ᴦ(ku ^{s*RPX5!HN"VqQ{UpI9eޒ{=Hj	ƭRrT" +;뒊(mhM7DP+42!K
99V4idDT/ہ9FDݜX"T/aGȉC	.ztIIй3~zZL︘Xo4g)5'A2+h=PPc8u2*PR1̷j!Kzkdݕ|7אn0x;l08Gs'\eo}vmz#jq;~%	D1b6!xjq˘ԷF$AH=G+:P^/TuW!]xh(~Rby/C#'薎-G%đv(΢TNE;|ʟ)o
Jc;htgYNގ9vj
MBy7L'G)L]Z\=XyغBG|܎4X;
LR]h0A;UFڡ֢3lvc޺|(Lfۤ'eK.w5(홈6f_+L=sgYr
ۨ3`TB+=	&/(pRNiaLx5c?i
kW x?%4eZYŢWo0JEAA(s#A]($5Bz}*'qzwە)퍽R<\:8qL+	|UG['}rynHrd`8aEvwm`sD,U)Hä!/:P#k?{ҔSęh^$|&9DlPlT@\DCX܎}e8_tbn_knWNsB8//ۛ/?~kD>aowe
2$
(O
]mo7+Y-"]	xe
/E[9rw
o՗IcCu(BPPEFtXBRCǠOi#	I	mcn(YS.:rds
1z	@R"hPb:x-~pfЭFC~5o#NB5d @Fs'P=P@uhd4T%6N@ɸ'2iWf#d"$vC;F#p6Jo'U&KiRC_|WbFS&}5E_X&g}\y<xRO|~6ʷorPkUHe*&<$)2$6ȵ~PdO2fdm!uXr.1mkp="3*ʮHJBa#26햱VifIƶ40{kH2;xd̂iO.|<>ξrcu*0^	%t."@@$R/9E,0Fm51*
VAtk|Q	->7ͺsŎ'|bn&b`{ě*%X4ya*AE|JrAYɫA"3lm+dc1fȌEIJmM&"e6ɂA#86'Kjw"6m-"68X"$!&o1;%60V	:Ja̤N1%rmqy'̍ɀX:#RL19 KV&3iJHEl֝
vqKkͤd[8E)
ZGY!bHBPG0ȴ1	},Ȃ`.Nl%[Mރ	%ELa`?>zMn59?z{^zŸ[/إE3+42j՚g=ҝ?=u>Pa+bs)&]K:v>1jXsdA8hn3]x&<~AZ\KK<PI.gCcx+y~LTς(fSɂ\3 (TU1ؔT1+#-׍Yw6Գ^
Ԭԓ7	*~]D@
 AO	A*)~1ʚ%;%
Dʱ*R~Za7ӏ,1/҉e[2bւ1R 2BK[ϵLnZ͠VNk4#VEeJ+,t29`tA}644\ᾝMwS3V|c[zl>"?,
[~"9PgbQуMBdQGlzLԣzn|l%![jmcqollZ'+KnxK:@,gq6aSiilta4ٿJd2iDIYC*JoȀۡ78Kk
"mQ6HZSwBUZi6(ȣhJmL8ee b&䁫4'S;L62y",ًܺElŰ;
>y\c8IJɐ&i,!eJdu7hp8YDkXqkO%Y+OeJѬ
-)DKt
!aA$F}6U!;ѹ/^XpҘi̗~;O6vH%N\tE6(b^4おSJ,Qb}p.??=ajziERNkdUW€)IJ
FQa]+yցIA*V%b'1Ёc `m1$lbf$D1_Sl^R¢EQ}eUV!NѿV,V=<շkX/׬',8e@k3]g~iXǩW/A{^%?Oq~hoF̠ƹ8k>Zc"Pd;{oj-u)}{{tŹq;J}a joVc)~uj=NCoD+|iq
>_>ԁRV/]FWqhR.Ą~;y]<9QѕgZ\=_^cD̫ċٻŁ51VUg>?j1GFgՈ߾d[
1Z2rmK`oúffeayJi|':u?/zzч
ҭmU[]>V=cV5:R6,>;z)Nʨ}'G'`MM7,3x~zN߱8/{}~߿>_/<3J<){.cV]YQZɴ~34L]W/S:V=Q^~ütX;0rMĶgM叛7De&MC˧^&bG^>m0%CMSoteG~0ma~_w߾]E_đS8G"g+`cc%dF)JQn!萊K4[G8C/6,F,
Žh=f0Z$AGʒ|Cu)R<-ՍQeadOg^WΓw䡫5jpyKθ>uƼ~<!\ LW76h3z<$7ݶD~K+QM&zbQ`da`0i	N$5I$c`LHv^x2S5R>[	Rq6s%2J	8YwnIݡM
e~V~<ԚX$z=ԫ?GW{8GBG_u0RH8r#wjLO	sp!^ŇOxvg2}vմfgO*Mz|tSn@Inf5WV xe?Px뫛_%G{;pQzܰL7xZ
VWw@y*{ѷ;2.
`g++98˿M0U8?6*O{롛Kv?+y[-/Vm+XaC:_|~i/Woś<4û͛5|qW~Xo{^c?ABָf+e4}|d3E9:	>:FɍW}\SIpG*θzjkrѩp+JDǕ
f긂JԌĕԵ#\`:{%:D%ӌ1Qw+&%Wq%*wf\j/_OlLy]:_|3V.<_7Y=th
-_\\ݭ/K43v:\3omΧ^nކ0ry?Q%v4f'jiIs|6t+65\ZSǕd?qbpVnp%rM7I~NWޒ=ى`gJ~1TN̸:\묣pd~p%ru7c櫟;$&vgA+#޴WaS뗖:[6._ݫ.Kvć]j,:_ag6nrMԤ4ڶ|!/	8pkQlvk#|a&l%5ɾ֋tk7۲[9Y/n/飗_-єgum2;&dUiuIK2){Ns2qnC]o۲(gx%1aiTi0gnnO۲ŶcV#Y2`mpp-ڄ2'Psmm8SZd=O):ZPzCesďeVx@vFWH*1vչ֝99>5u
-@1˪QxCԖCehQxFb`\OqZW'+y s?bDn?P:DW"<{Q',FTyq:•
Dnf
j;D%ͫNWZ~GhD.^p%j?.*<=L9e&x2N98|GF=RpWOmzpnp%rWP:D%W'+1DpW9h8dc=5N%W'++D0S7\cSŒkvrb΍:9KDhQn8h:4Ϙ:zcZTӧi>8u\\\C;{gN:DmTNn׸WW*ϱ'p%r=+QqKg\q;M֩`_Wh55/wqy)%	E[m.s߿1
|?澾la~}5EJ,kXs(_]=cz_^?򪤻/B.Z~[]߮@nƲ;α۫_lOky{T.~sw;z
aqp޹7]z{?sxcSn^KH?Zjuo~|x'vU~P}qf9:*@V+	ղNpY
>y?\5OO|F|C\FAwoz[
۶/'>SKȖd;ϱ%a%]&k#;RLNgvE'b;^/wI1>Q{^]¹ݴC6,Q4Mm??̐tQ*-C7{cd8luR5FXWuftP32H՘FTT2XrT2ªyTzs*[[C(4ОrVmP"DeY
LT>k˭91*U1heC`&UZQ37mPi:
9EJfq{ʆ#XC.W_߿AHϖR!Vldm)#sMsjJ
1O´j{-}LBu841`C͈XsZC))w3$|Ģ5#񵾞ouOu0h(sewCj}!yXj7EעְS#NE[RHppN?AϧJFh)V\@5Q#dXH,!8d"\mF#QU+I;MJtAQrM$:=>5sS"3HTU]PY`BMLM#|m.5~;0Q_fV
]dbgOYA":[;X4  7?pTj('έZDyU벫RYdPBge`5ѷEѕO,:*d=(
$B2zCL:m]=>s%psM{i@0O0.^z8ڬ7uY.CQ)J5odJ'()v@=[ll%EMjR(c9(+H\K<!VQIiV2LBoȱl&) t؈ZvE}˺2|3B5DoȻB+X"cL҆"1ԇK$A(fƁxp̶Ut&#ox!^Tё;hC5@ pc_JL
r|BQkUa=Dq$YiAC5lBD	GC!EVS|Кe+{@ʆ0BZ()g$ސ]eW.{^tY"%f4eVJ
׭Td!Ãz^n!G%w1BH/gNΓkeoha8HRlBX/Kr881,yWp&мwY+cКeM 3(72~RQTUE189b>U86vNZ#"_	0klq¶,eXs/d<|Q@z|oi($|ЈmA-$4>*%uPmD~Xw2*Eβc\I!Vda`8"	'bG:.cP<I&|Zex^Sdf
Oe6:,d!y		$>yT_6	!kЉ( "[
G҇"dN5DF5%awV566dF^
Hl5w7yoRM𓥩d}E!`d>bE\ n,zKSv4ys 7RDcaL @ H~@,}P/a3}Jhvر, AWBkFO,K=hhEeGV!E΃As9&Ađ5;1vA!`&֪((ͮĄ#<(Zrw7;k*rVXd&4!gZ{ȱ_]6E,6f;N1`XErKgdX˖bv]!yyx/C6,0k09YXà`
'ORE
ޚ4^K3OQI 2ohmJUJ{(f'Xp-on019`YBykIJĐN9i
B=Hc=ŭN:.'U,<Ӫ]N泼\Ix[`kE܀-d0L[0S}m'H;Xkf!:C(p58EC>v
3J~6{u3|On(!zM.):AoP<@ᡅdfh+(!kh2BO^PRD5HOp'xʍ׃}oD
z=VWt"y>5\Apo 7ZBs`@"m@(`^XJ0в	BZ&@\v!J|R	BO連 )C'I%n8낗`A \4
j٫͆(Få#Dh``.zǚYY	$o!HjvSXrߝ\Z@^Fu(?+,;70&UGR/\|oj]VtU]3^ڋPirt q6cyW: x;.

l6VμYYa .AkrX],rg"Njrrqϳ"	^,BCܼܮ[j;kAw"|{~oӧ϶p|GWo<ڬa:Ņs}6-DN:ٚ:=&~Nc=ž0bSL/h
Տ-1tM2/&Uwf~ܽԮmw‗ oTjtcj.k)(~́v,X_!^I5bM(f
Dc.t$:JPW$"	HpE+\W$"	HpE+\W$"	HpE+\W$"	HpE+\W$"	HpE+\W$"	HpE+\=JpKuI+^#CqLjRz\
2kIpE+\W$"	HpE+\W$"	HpE+\W$"	HpE+\W$"	HpE+\W$"	HpE+\W$"	Tp^%	Ы7\Yo:;\
쳪\^]!`ˋ+l)th嬈]NWU1 7z>w=0]Bk?]CW=J]=9DWb!+`C+D94Е\[[]`MCWW]!ZcNWҒwut%DWX*U]!\K+@
Qvv#:RLj}3zL%k}T?6x>y1K'PP(ͺ4	ofjYo+HM^UXq3e(2o7[^"65x^fycRyd]믞ghURʺRʇT!LhRJg0S8Nf
弜Ou0/UjҾzޚ71Y;d0|u*ɫUS|UWdJ6Gdg%k̕F
'yUZOGYDOLmk>Yc]fQ[:{ܲQƫ;ﻋ)xی˫DD7Çeqfz
ddwN&i~tx}
[/5_~nEW|ǜv=&%t1|6wfb^gb:m~38P?/Yf|z/;]_OhÕۇF7avr*D4N0OSNbx|ovn>&Fd0wzcoi8@1Fw7\w|]g;=
t舧M~◆9KsR	Ӣ!ew9JSLIyO*ehM>Q#SqImdZ0L맸d6Ymݛ,w6JnyZh:^ ޟhaq9F߆BnǍժ_ͫ[=NCNv	{人c;o}X+.a?B0=-!{?Gǵpg,\
mB\Z8}Bj2ӵԹVQ0^Gm!gᘫ98N1ԵJe.
G^kckxsW){@j?~׋i9 rYSy])&jEvN6`J稒*lTKYKcӘ^jƓUZ@c֍[O+PPv
jv뢾Ёx&ʕ:a=9eK"6yWq-AzՁpquHkմdW\.bL 3"[r-|d*_@12
XgȄ.O;jG< lyᾦ~ȭobGy̝כX
{+ыU,.mO8 !u5gяv%/Wj*=*FnޙV[nE<az6woAO>p2>;ɛ̑Ơ~2JF0kU&JɨN(_"8ʲ2dLޒ.8ONMJt&|肅}GEgq.>\|z[YG˭#@h@ԑ%T	歑sβDId%F5 ^0r ![4($,搬:)n=(knjM(|D:pA
.0|cC
>&(^Mɓ'Mltoj>n?er]X6 $mi%k*]	`ɰ.CEd'4@ZkC^0ITFe, yBpD䩂<];	Vs8FodѓNBxHIX׎@@OZLljV>cZ5mINQmfwH
?AJR
/7hg_fep,
ΤeJs/nmv>#cN3;.i휀MS0~jJ~52dtBmP2jiW>OhݍԬ7xB'7y}m%\5dkTyF-vOt:ЇJٯ#w6}h#xk~Ox:o^ILz9HR)U;\5QHd4Ξ^(2CC=Ķ	QL%!UY8ex*CC$MZut|prܐQA酏tǥN5O/܈J ;L+򵝤Q~\>|!|tqRG~[eo8uÿLen*!|W_NZnHFs=/k4׃w	'/A{	KGq?7;tY;۬R<̣]a؉O7%ygzߗ4
N?TX~9LSFtKp?p]ut7=%K;O(1ιQlO*C']̓3޷_J?=y,7C1IéW:,:~9O{%㳫?u%ĵJI*]06$5{:,8ί]e'Ypס9\"os=ht)ag2cڅy^7^jͅG<&eoeMV(`10dVgwKB诡;9mX林v-œC^pRybitK"X|*~&GJGymXwh2)Kw߽+?_/3
)]3~O'0~LRr_~I
\'ϒR)5]t&I&`,Hκ98K?vws{.t9WBD¾?Mo;|g惼"t簾0My6,"gy?}];?-xQ<>3O{Lh~xS3[~鹡d5rr3_.O޹oW/5yY>=J[F!ε(hUuZPx+Ui^`k`[r="@IbWgQnQ1%r"99{`ȋJrI[1|ly|KL]Ts]1a0xuv+1ZuI|$lK
۬2d;Ⱥeb*'lΠ8GEׅ8f@dT9ڵT݂;dzfޛakSb9U[fZ\jAze
$l{6иN#'n,%W(:.%]!d.K5OKǣ!,{m%R-rn)iiQoOVo:?j6KĐ) jcOQ$L0c^rtC
,\_#)Zڠuv9?$cv;opSU;o-w}%
B8aRqR;v%$6+.Z|)[٘AK98Vıؘ8ƶ,jSkV+v"Ab+:8		,dZDKu(UI5Yf>DuҵA)#QBQ	C&ՃM^%4lQgcP'ȬVV+ޢ11g<ToGCʺ|rdLV"L2Lѣ._LY|vA.C.ԭLR[G$TcI*+4E=ry9͛ௗ4O;}ͭԷn-QΏݍ~XukmH
_Av~?wY{-paSF"#߯j%R8	'&_T]U]]u(p_|?8[ns
_g@{B$ڞuݰ%9_fYX$1<4]|88:ѢEap2 nF1uwIl`ëM_Ro{H/	
'M
yQQ2vX0A:Տ???Ӈ^@o>qɺqI9P[0zdrA?
xLvE|Ag	}@8	=Orcl?-w$y []ǫ_zRY܈mMz:~گzA%"b{<yd~PdPc~CWMmb'>n^=Qyv.iM#RєGu 1"@iV;O`F-7c5~ndɇYpkq$rbBoI8`F[3mwov KMRMœU^R6uee2KtiShSKϔmnmpۨrQQ))Fb\*Hd'a&D"*kB굉*l®2>?k*,9O7u<:3dtiMj&IMpRJ+ZyLE)0$̏Sv@{FҼ:E2r&KM=n@)cNS$g,:'/\͎H	F:i>Rl&YP<2уZAa$Esbcpra:qHKYэ|(/}rc`UJmR*RMf⵬tU0&l~9OOcـJH7*+AL'8)I:5Tr5KM-**+jHpĊ(DB*j2BiQ,\J"v1KEjNɥdfoM)zAyDY&1Z)ϑJUjjcpH|J	R¼H4*lifo5	閞FW>!oaf{s]xVl>l:817*6(G-
Sop\*חgW.()U26!IJTXfw@Ž@/nXz2㇣7p[&dj4e.f0pwBO	A'ϣэ'ѡ1eQ$_U:ׯS̲035NS
eJI54
fZE}{*w=cZD
7ć1Nk^~hIu5'}L﮾KFl
hwz-h.WϿV$\S֖M/-S:+-ȍFhKf榱
,A
dvT<]ja:{?si٬DMuf7MﭖyB~ӳp5nhofb<vl>=
O6#|zChΆ?{t\>#+иgkӐ8Gb̳5bRD&5IuJYQTs>m[)Y)j>T	(2Î9)F6M@ft;+jT#
8
l&N=ik1>">EҲsٲ]jl =S|ôs9B^dr:ʩs %%V0,[c-ZcWi^kG2Πxw&Z	8pr
R+2ezo}pFV[Q҉(pBZBW+Dɻ#c+%6BGugCWo99!D(~j;DtG0Q,q'َ)_DWXb
jR
]!ZKNWŅwll1tp))-m+DUGWGHWrcHAt-/Jж#+ƿإ-5M6XdouRz-l
6vML%MdD--I]JZIo1Zv^-#a12؈ض؀R֙hb3[WOk
Sø%#h[iwPiDZOӈҐf>trΙ-k}]X1ҕf2S]!`+k153\0ЕTX^]`)!r|vuTKW@@]}ËQB0FjЯp=^295LJ-&{1@O݋_\9:o:G)&Zi݋Mau?n6=&̍0^7YOhio¸wk /^/?Hq!2H^Hrsqy>N䨖D_h6?~^EGC\Wm伷z$L-̽ ̄Jhm9R_`oOz'䱴Nޤ)vstA5YiP}{J	]?˅b%"`Sq	p5)&_Cvڳ薯?{׶GE`Ӛ<	0`bO+ȫE,%$Gc&EqF*ELWM7PdwuWd9q"2*_jt5?J}H^~m'vU\IqD*>4&#t.VD-Zc͘hS7#`sﳻ]x忍k雗{h8y}yq~go_2]׋H/}af+~|oOt{W5?<2kw>^~^|!jJtҮ{5Ǿ@9wxҟ_}tO9YU'ܓr+1[5+OyvƔF!~Cfe0ϿdhQ!X&
t6chڸ-d]!]9H
CWlF]
ήpc+W]p0ۙ.pYBW^]
4]|t
%lg
pl_\ӥGIWY[+V]
BW@i݉~7tOz~e\0.f-r{=V硴+{
Ϡ+>s<8nm#v(fOt{?m~n3t5
]
άJ':B¦ GVjU];]
Ԉq};9j@6aǭ5]j=z{5oi7⧭ָi8W欑SnsYlO;9nr';Pd_3VJuFtSogf,GbJOSG8`7DWyp3z'yhjDWGHW.8-]3G].hYNW1ҕ1l38ͬ3hR(OW%](>[jvۡoF]
NWe':BB%uVjntDW'"c~}?=m\<@7CW4AWr=16DWp].
]
GbJwc+!_m|ep\%gBW@+z(Jɛ0?0nC]CWohOtutX+L8.hp+`ЗCi4=P=Ҵc9;fΜ݀63g7vJ1'|t
f
pɄ@+7!~߄@[cۡ7nF]mիRN*Z28l3~+t5ЊJkNt+}+j`o":?³s'QC+[t'zxV]5v3t5*oZk+tNtutvulة]s~t.~;ylwO
CvD81߽h1=iH/޼7oCp!t}P7yԈ
}"}QIgdcz|7K%]|>7|Dw[42ﮮ/.,wz'9ww}r/g7JR9i3(!~O_kEJ}גt_>wpwF틷Ȕ;~qӚ#BH?w%TջO܋cOewo>-,WG|Fpz?[&pBx^蛯ρY)O@1gL???iUaQ_`mb^k$7rAjT9Qs,݅DP[!q|Wg-${*{}@{
eo~;4uq;{_>'*DFmgM
WivR25d#jc
׳jsJ1{BR*"&U#ΰ_K9W.UݛRy_c(Lu6Yq96wBvqV0cΑbՎ*n2LV[-hsRtOpP$ZZj
(V1Jk`DRmhFJ%E&юK1%Fѹ\.޽=$T,nmn
Zd%CI5F1=F$BK'cHfcgvh춸\蔴h1-^rJx <ф`֨ߥ˿穴7R!.V^yü9YrX\O/Ss.>7'!ȪB޴qM%%i(%%Qh.1HGʉ
7@xב)%Dk(5Z%GYGgfFK€bkiG$XrxQR!hEls!j4U'"ȑ,Z"AM)PA*n\3>Eਏ]j,d)	3ȸT'Ht=K:0AF׆R3uiJlGxH(S`ձ*˒`gg,a{eؠA[wZy9uԡ휴i,k4
%Is*j]vut%N@2h˳P
XBՍDWX$^.CqkЬ
0+55PB7l>,	֡s6c'3P(kGU((MUJ@hK)O-`
t[HC2"Ȕc!]Mh(pu
8R).	d>
4׀R%꒡4LR!zDzgb2	$@@YVKJ+%TR

++Uq2M!P !-	V $@	mƮvjBu>9÷	:	9q&	cBl mKI!0ά'D3
2I_ fU5Ttg@Qā.(ʨq`j
^Y5(loaH=ݻ+Q7a^TuA"FZ.
{Fd^n!G3)$$W&PLJD^l,@v1!
0G{xWZM54-d36>r_ݢŌ41c58	4p	U46oʐvH'"!pBDE;OvיxaWn1K3'/x
Z|otڦB&a-A7
/Mj#TYǬd
5{H\!	*#d`1OѰMʌJa!)a!/Q[tClM1WTs]`7"2UP[h/&%\I"ݥRAXa)Ð*PČ*mAz>
y[
3lb
VFqQvRƐpM9E~$oQ^1b^8hcyآR47QÌĢ;$zށ U@:\biA2<4xX)ltZHhՃ*LJ6L1Kc3Bq!;UuԽtw"+뒪:fc% {[Ozb|(b|RF/LF$f3~h]M]TӔʹ>t~l˘~\(i.V{xpX0pq|
KjfrZ?:OM&t'^ftfSc[y3jsJ{BZ9hן>QG*I	=#\E@i)NQ	_&%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ R@"%)H	DJ RUy~HJ p@׺(|浫Кg8gotGDW?Rqp
]!ZkNWR0Ftut+W
.]n(th;]!JM)ҕF9 "\9
jwB$])^k6?c0_	g:D7nH4
i4Q4hMiQ:SiÄC
qE.CW펏yŀhe
Qe&:L~tNp=
]bBrDW'HW38Ԁ
;1B~0 oD :EWp8whm+Diۡ+]o`>;]qp3B+8g#s]v=?t+P
tJ'~sXB~8ts>BRDW'HWGu+	.L0h;]!JeNS^pdeX8f*&~9wpTςPpS:+[wӝibA^v/![쟤˶؉Vg\s0IsXuLCZ	ϿAQppszJiii%CWWt(@%z2VYDW

Nut(Ŝ"]Y̐uBC+D{(%':Ery=$r0]\>+DL
QZ:ENI)DWX$#\=
]};tes秫[qp{/:LoC){}{]Yv=n݀
wPf0tp
]!Z{BN] ]	XB?
0gl0tp
]!ZcNW3+-?n}ЏCU
QKtuta{iL=-|ń3vȃG[wnL5*;'@ˋ5[_UJRSaVAnʿFRvM<ߖſ^zxl'ij9/U.!VJ_]9Yqڋz?%ߔo0{}	4ލEl{Zϋ)
釜]x>{)t9<緶wV)&lr{6	W[̻sG_G{+D>r1m}槶oos?׮Vx:*SJX"ZyfKT}Rsε)v8ovyohi	TVJjՕ:cX yLl!gpՎS
Rs2m$J^Wgvnp곔+lIv9Zfp(3lkV87߾vur>l5ݿZ.gr5?? 7|j~qw
Lgܸ3.tGگ3lʞ}S<.d";'hq\U^`E:Tb1=D%Huf0`M]ܜ$i]o//x0jmQd`h#S0k]xg^p%TGxݺ0ْRkwR(]tڦkBNlN8* 383cwJg\ؙe<u\*aݚKwCZb*M/vuluTa6[MW7:yoR2X_	E\
3jpDjolU^.)#1;2-1"1,>y6
]E;)w2AP' N;Θ
bm1d(3uSdí.Frkw3bgpgw^\Uէά䡼S5Cajme`!plCd,Ge LT6(/k/~
^썟ؕu<3>:C0>
ˆo9A8߰:&K`}vG^t+*{xb]OࠡO~H<"BS-s6q1f.śQ[w1qLL,&K17
_U{x|
jf Ll轆?4^\gBMPs3
)԰w=@!ۉ}x¯2(yj^%Ÿ`ihAңbv:;ڗʕ+=+5%=f}դMkѻis>f4LF܎p3mVh4rPm}'p|CaΩQqQP"\+ubE9T>S268'vuI=tڲi{9n[6.Dž7(c+,c{%s|<9fOn/A{	KxGlwd7oeeśӬwvը6x>8>Og	;_\6FHM[_G9LSnѫ2oJ~WbFe&6!hwJD%9+Z.w73]?	Fcgҥ0_v9y}qdo
Oh8<7\̄D'\\eTv4W+WF?<;
M;..Ωzy>4OZdt\|@zr2C)Lf`bg9EM>Z)vLVV
f~sIf3Ի^nSڂ%,WR*,/Yd<
LF7T1O,J.
nowe](,<`eNFdN%̀9FPa9]T)I*D@`	N.%lz5ݺrӪ}Nӟ(J%
=z~g+OeBsV2(.bI6GHCe	{~gaC*sv{1{<"w=*
$ld2n	{{y#\]uq0G$Xxue/uw589Ts	0FD:rRdqLI&d4)m)ɁCbP`ٺ~&Wa^HqPwaڭV`j~(݅ne+-fNwa8/r`L2Xfl&HÙR&l';.=OVX.9ŎYmdˁdhg>GP,MʶehQ{M\{wڎ4{Oq;r!_B&6;>fs&d瓑t<L\f5+~zu: mUwlӴ&NZ)=f|ۛc^ц?~3O
ΨJʍ5Aה.!*m;k7OzJEZ;ѱcysyeWbg7Unߋz/$I4{2QoD<	^#AU㼵RL7oQd?E/엵eJv]P:B(cJs4D)c)L&Y[UaȃZ#-c<ˀ}s:I^QlM
sH(|9vZ.v;#V֚()}-br>2}>g[ݪ	9IGJbZg3lHVh3V6ʳD2}%ĂrHH?H[$&\!#O"R;yB'#O"zhw-(AUE"5RjX:R&lQwZB'vRb6|^<`%=ᅤL}^H[]Iɬ'ٯf
?
8"kQ_7sl:._a؋:OAWy]t%ν^Csk;AˠWτA+!3I@gKU@-
5D{Nc+m0MVS<=L}.a"6gXlm^|!JK}D!5(hTqZPx+UI^^S	
%'Mr@H3:$S,'+82 UusݥHQ,7Vjp񮻡۞Ԙ&T+/݆_/ȣ-Q'tΐ0$d[R0'1E-Pxt\`w9=J-#rCQ،*G+cpZD.cW*c݁S/:brD).'UJd"2	G쑛EpGhG㤔YaJ{Ojc\:EgǁyKreI	TX5sLڔ4.10'9J\U<,{Y1NLr 0pD~L"kZKdxVm;[YWh Ps6(jsa	EN$0&'"
5{~-?'[RHI"7>ۄc"u9gHT6
8bZqzIDާ*G+2Db9"Ӥ1Y
joHG$U2kBE120H%t'CO8	N.I6:!F,0q5[Wz"E[պ%?,oæSLq+Ӓ$V`thr$q#Vsr)%6%|^d_|2-@s&B^3y	%EpD&X}4A JDztѣz,&/;AȎءY]A&FvV@ܚ4f>䦵-ݥ1t Q"	}lIlƻɊǷ5
֨uc

|}J=mݣ'tՓb@RFb$Τ=ziw$$#dtJ&D#UcI*+j*xN0B; F8cIՖJkmgˤ%5#r߬#{|EL%RQ"LZD4ҋiY$9J؃9RĊiX_,E]
-sT`m{S:#&%0Ch.<1R|l#pRH9+d}yk+weeh5.Jo$0~*|ɅǓ_p:!OM;M	5'%%,s2wS?J&?`n)F7FyؑZ6AAC2~@T(]D0gR2q;FJ.㑟L<]MDϏ5gקrt"c:?,߮Y[tM6>Ѵls~y?}4tC['\#b''[%%Fb	GqQb{GuM\/W'N}1M(x?ulx1я'o_^lF.9
Gltvqd0~t_-/G)ƺVҼs%ׯt޵\V26h'rGg׋ܬ9<N7\Uͫ.ٵjY-ɍjꤤGše,S_Mp0΃B6gzӤ`ɆA^8ч7MՇ^y^xq?
L1~]zV%-lO&W?'$^dZStߐ3RU<AZ|N@>ˊNq>EŸG_߷5x-?neKKYZ<·Yjck5ܐxZb‘,H|:^2	ᩝ˅xvw_nԏnK_#z:Z	$W&$}"?z|"IycqƢR:seMa^S05i0Rȕ82
JGw 23TUk5J^w|ΝC`5bU2a;JK}F?w|қkEXk8J%fMH #DPcsQRӛwjd✒fSyv<_>[Du(J Y\]^t0f]4
Mo`qF&X[pLs]όٝmbO7Y?_<ɧŻ<g>ebuUݭħܖ1{W<H5߮jeU+Q+Ǯ]]..
36_2x5gz}_ZDž:9X[|I	mlhCZ%9Lk"5V̨ڣ<\X-1S+gA*C0TMuiPA6蔵P$:]c5))iqˬR^1c1:Ӂ"ohTk9@wZ9א%3L^>8b5etRE[Îar{鏌)ܟO)}/{èт9t/:L	WaW҅}AbUӯita
vlS/i?O~}ù_Иyers[b>̦[I(јdPv	[|_A9BoD%Ǔ],k@Ե}>_VI)4zE7L4fql,x
Z>$e)iMvr׋;?XNCn|_~n|_6']8Zhom>=<ަ|_;\pl	}HHWtҷ7[y+hs
U7i6-Mk4]6=CwAntt<ǷJr&:0^][@bj]x<-[~k!#{@~M?LiS8QZ{nTW.n3x)̮"@H7s鄓O{''m?\wx+񣿡]"s7A{4O<~?GҮ4qѓi&~eS@Y?޹trtq|s@M{>wЛCzʷݮs׆SNΛٟN]Z:<kn367|e&JF52xum$jeRxޭK%/2u삵8$>wL!	E@:#6%X	][:%(ӣL
;p	RGJ>#EW;ĨRj`.*Pq: Zw0aoʨi{yVX7@X=礵~|um9b۩N KR>y\
6y2
|.)KĝI2juO\(f,ʭTmUZNkk~gP]iqoW}q 7`ځHl)8.R	DRΫR"FBW7ඤJƃ1skS$4/SQF
ղ%¥IaZV0V'C;&ޤ	^iE0\YFNq͢n6t(o-g&)Y5.Z',cg;om$6`#3tTDr]86yoQ1ZU*ڨF\rc'ݨg,yƯ^a쭲8ygpzpK^/̒:i&]?[+KSPYWcMQ|9Z䒏`Hnؿ휁rgcnǜefDUaŖY&x1ańD)aDk b5Ê
'W,]+"Zr+rU\ޞ0ftr5qj-0J!+;@lv=5 +V"\mK+&w"UF8o}劀u9ZY\!a6w"J^c+`P\J
\rUP$]'voH^O)ިh1id6M}p?yX'Td.nl+_|vVmBE]},fc(z^а|0n'W4}"	1L//n6lT"tΏI'ldQ4Bm'ާ;%ڝO=ẗUU-4	d\wMAroݫy]ˤzvjg{֚«Nޥ4\N7B6
ME{3jӷg7{}b/'3u?2g?|3-F|&nQ
9z3pjeU۳hJZ@`šbu2ȕ۱)NHVF:FkF52-]n\*WOzP%'Wl-Fw@aZ.WDit#+4%W̹*FWRh]R@Qrۂ
5bp(EVCrEU:B8W&h~&PQԭڭw F`L12MhUV%Q"ZeEdZi-]IV%++uBZ(rur
+WB)rE:{(sr"reW$2}b
q5S}vDiD#+k45D`	@\l)rE2K`ʕ3NIW\!n9A2HڿlY!s+=r@\;dpq\)3|$rFo*l*4`LSV`ChLA6c+y)@J+k]k$W\
*V\qR.U:B2L֙Y\.ƺ"Z
RZrure[R1s	b֊
)cUPS%#04$\)J+5*W_\]7xdqȾab)U^%+rԮҺ
_bp*EV3]*WG(W;[\!.@1r\1ʕZ;V\!)Ⱥ"\K+u:wB͘*W#W&X&qTJ-i3$C4ZV$h*R*G(Jp%i5Zfr+WJ%/HX#Wl)r+UQp&N$We9AUL֨(3K!^eJVDr+XWD[+tuEIՃr
q"J"Q_\2i>ظvµ/F;^A.>@xv=(7/Hk]\
&s+UQ\
bp+Ev-(F
*W#WBYeArʱWc]6Q斉ՋȕdBo]e^dPp;f9PXbuAyA{2%>KuZȲPzYi[{z;C3l2L?0ެ:䶯ze1
Z\ 80A9$;RA#WCĘSm}GlZMo|ҿZ>}
U;念vͬ&YZ߽_^\Vlۏ_6H(=һ]-x߁c'_[=RVO`& \r%[Kܻ9$w\L#a>
fyd_#eӝxee]!x=Ȧ8h#?~]7|i&-WR]>?o~[?r2Zs=镈"/ߦ5)Q)TѬgd6؄03jԔгzz(aȑZ*Ÿ{0)c	6h?
MρǧW!DE^pTOW(^u,1W#X۳L8cmXJQĥBy[Oڡl>J&3E'g3/)"4GwK:ov61Bfw^P}
M%LdK{07B"a'r;ƨgҏm.˞kP~Ufk20qo>fL{2LO+C $LW[hh|:,JV'P0G֧W޹!CB%)Md,cHhGBfRJlȄK^U\ܮRE(D
"1(aH(p0FD&F*CQ|xӪʉԣOO2Kw|2Z $TR8iFp5I㹞<֦fJȃ:ln	[6ս$0P]!Qn`hJXQm&$ubZek:)R]-5.>ĬFkmN`YdyC⫷~0-5,=QbCwƓ0b1\$CG_qu_ZNyT~>0]8?f]gO5c$X73yAmN{b7R"D8`8	eBoiz7в.>/.ͶLMS}Mgk9VXSQDY"CRY\B!*"\A-07Oi8{?ykF7iLIٔau
n͵
;?\~(2&~|-\&iUPTஜsi'0ᦤupQRdciS˴h7&xwbI2݉d;ܒ\Tcŝ	ɤȕkw~b8N&)N9`1,K8,ԗi^=pDyaG06$v2&\BZlntJޢ۔ڍYY4'gb}Vh).՘nKsCi4RHf`j3/H_iInPIJRPw`B'-	L3#wIzβ%oS֎M9H/~aOv͹ޠvN$K&#1[`o.Y?	1^0Ikʒ]$+_dm97ҩy;5oqacVidp2isӍ.Hz*0!-T	`Rb8qDŽ_XO:HzWjdMt_<;(kp7\8m1nrԘm_N̦LN+_2C9K.v;D'i͌F[#c/HV#x})~I
IQtUM9n7.欼!$gemo/qUYА2`TSVS'0T;N
IA*eayvKVXI7j+61p!anwÉN@#C+U'H\2}j$($(a슋݀PJH$0)Cx=Ji!R	Q&#H2=t+iXIc"'`ЭX5\sB@jJB2\'9u6&ĬBwB'ǸM61$
Y!e];'00zyv.vJ=4dLH6ƛqbV֛ya!ɚ$euSuAP	$$f>Q܅3uߞ͛Āe36S^&;{_)?[H7f#:8w[Mo߷g0K@JA%˨^ #&}bF䀙egJݡh.˲.vx_xA
FlK,IXÌu)K+.c~}ۅ^/}4'!":ROpq?^vw/Q	./s{BQbD 4fAY8Y?N#ǚ`^PV9%% b=5Ch2\؊U3J>ZbpiCn]7InZ!z۹7+K~)ZMs/:V^uĔDPk
p^+N5`^-0lh+/@όϓV?M&(M()S.7{+":t;ݤC
"tpÿBZ_E]V^0J4.}-WlBkATW~~z#\HT"3RH*@W(wWEv"P>L&t،soR(M@}_+Z#5!U,y';wD,T1%{X5ZfJ|F^rARo;DRz@k"^ͺ5FzJq,XU6Gn	淴7L$].!wVs>~ų<(Հ9YK*	qLf KK
[ۜ82b8~A\pYY(o|u \w\t?H,*Kzt䦊ՓrD5zY"_L}Vv~a"'ɵҪyo
2dx{W&b_w9՗dTY0Q"ʮ`
WksTZke'W< T{~ܩm;$i&5y	
*v b۪L@
|3eY/yr̐ȗ
FǟhDp.52US~U%UK\+.bd;2 nP2;["D42φ;f8*H_*+_~	f{ b\J%Đ&>jWߝRUW]Dž$^@Ll>9qrΑ6\G̨(xWئrpvڻUtkgG&|ay0p->
:Nر/DzcŔIʎ:qJź>	́M<ȩXﶬE$YVTM28FbVEƇ-g|KFIb1z*kz$.shg}|uQJDvVD:+sLU#}/d"@w4&d-g}Wkց[Լ@*M>W.M07b$FUdj}MV'1zAHqy"Ar@M/A`
+V(i$ 4e193*KE"T/h0o`\ʌh{LaPMN).4S?Xu|(cOpqNGDNY_Эq1X[ 
\H{rUv3M|ã xiM-m*P:MHZҕ~VT2ج `LGy^GK>G4x$A#;2qxBѝZPC-R)!އ7h?@GJhniSbI=`{?Ub:H69->ӜM?݌%d'Ulh6J)[	^/#YOr't2dp.]W=oV*ϋ}ߘm`aLS;Bv'd'K/=DF]\EްdP0p"c"~@ƞ2rT,2u,m
^)kQ0G&Y=~
#Os[~@&M雗U<.>CiG-qP}H?}X
|[=f>=1	͚'Ts߂$IPJ}	rQ=8Qa1G@CEHc	dC
%^^X.Yi!``2.W6)~'}ַ~Sڔ
Ɂh@}
H1G}4/
97m+-s^C/rR~<	hAl4	iN

`LcJ<[KBYq&05ǻW-iPt}m`>{y̒p,&YL;3Q(naޔn^+/QBO&?* ±FR)D[G?bD\=p,F(OGs6E^H>D;C
+h൘("T4nطL+Kr!zLɻ!J&O-7>U7$7vUڪh|_HaMؾ)(y`$(KFfMi	UD3CDY4FvPmMtacһ>K2!#R
QwΓ,O&!ѧ(Tr9 z}i~syCtxhH˟&f}|E':ߔ	}GS?%"#bI,E),4i9cCO{Ѩ4ͳRD=GM
9Ek`xRRQC|$	&~%e)TRH=fh)Sk:$AEӺE?=$V91+YHHJ%64ZAB?OKԐDkԋ\i#{/
)`RAmmŴ&7u%$eEAAFIFyATiW
:FnNʋ(9ruZϼKEeN$9:&+RtlCNgt!Db%BY*s'rw9%;>H==,;2Doukse<^a|)\;)Xo'_KNJ#10|ז=fyAOɊVk'N
͗D%
_L:M
QJ)s)˔N٫T|չ僖E-m3IvJXCBjd}UFjb.!mM-c4yYd|m(czB%Ϙ/pp!r9)BkHHYg+~A|uhа븐fT=ĒĪG5rVO;e]s.TƊPC<"zL5G]+rCEB̬k
^4	O*hnᦧpA1&.Xv]N匯.޴>4J5S.zc|埓5z7657YL&FGLY7Ƞ"ѳZ1s<0&V}XuݕEt/i]#`$ppuq́3Y$97?&R~TL35|(771ܳ9TDϹ̓NQC}_3(?-g\'1J/\
>AŸ+Nڕ/*[,_sIĴqe)ew)٩o_8!yNƓ$Q(Ӣ%.m1/&CSjx]
N\w-&]PS=Pd|XvX~%f.hXH`:V<<p[׍(]0pw*g=c~E5&/|:$cl
8)1wq3[
Xw22k0gCIq%pcPb-o9	3Q)ת 96'587*Vհ[/ުb?Bn 0ET<qM'SðFP%8EAirLɤ#9WQkApu4w1+7I coUPy/4U+7tUu&D4Ndݟ-ҿg;h9ZKrǹ	"7$Zd/f帎vs#
~3N:bcipi^(fLk5@T7	4
.Y6'¾ZoVb@~˯U|y?$^0m2V7oGW.g}è
p"~pnWM2mJzoR@i\vKky5[AЩU>H(:k$o!nhQ*Xͼ1A8d:8@5]k{Z>'ʔ(e	ƐYy©PC^7Ԇ*ԝ;2fgdT
{ARvx)`L9ֺ3Vm*b4_mݏTf:%F艈Aq^&$1VpHkf<i~@ Eii­ZoK+JkmbtZzip2Uu1䰼m
@S@dyzk-HY kjMњJ/m-'Gy셁d;bh-fgq?5K.\`{VLPQ[>'T2%ӔTs?T4r\W译Taj<gWrGV4nFRcv^/UccۻJ9x΍
I\BDIŋ|=_Gզ+i/#.ӧY$t\U"߯o-oy/5WWׅuy\ZGޫ1`>yt4ͽ@55J52kc[򘧷K?PSm`CWOI̴RV1%9[kOG	_v+)7RBjIqbM\uaҀm5
@!NQ
/8Zm_V~DG#5!CgXZW׈P=.` oY:R)0>0#3AӳaߩC(:lhfǛ4^0︓@4쎳u~w/{)0,WL"'F2Up}HMIc|UÎ	u%,-LaMxDy5L}:VL$<;zQ0aU;ӗb-$c'r=92fltБOD
]SCO0p3mn~XEWGsIC	h;ʺ-O~EY\u
e:ԍ7F~IAݓ%&![bNʿi!/G9Ųb
6{_W"#(Np/I8'S0Yf6;ũ4Z
hښ,n9匁TV}rP;38S`b_^KHBhY^d-}=f_^I<~I ɥs~4þ.n;X|`(*$ȑ9){A|g4u^\s 	}GyĊɮp\
C|qs[6w{ezI?jҚΉ=Z"z2$;8i#zqkNԚ\c6aoafP~ege=euyJJC4JT9asz`!qK%D'0%oPAC?ȿm7tUbh?ČO3Kw(x	C܋"r=ьۯ}BPH0bk[Wϯ,nԒS)A
d;`8iay20bi2֜^167h"%wчy{	a(HwB(ߜVqzv5M+iʹ ^˟4ո[]%$e"+G$+K;]q^jqLk҆&\Jc͓D=t"nJe
F
20;e_|;bE)2~9U߹\Q!Yw͖棲7fǤ&C؆ծTQo[OL_xVQsҒsRMMc^qP-ldex)WvԢWmT>Td2ynIi4Le&Ŧp!:`%0U>Gc2$jIG|L
ldnJ%7)5/g}w\3:-d|Rz Łrճ>;k=ВP_ӖfKr
R}"׆Rk1-#0bi-:92	wC	X(΂:;iBz$ww"(uo)iҴۇ;f]9͛6Ie
0R7`ڞGTZGTX~*I5r9?fx/HnϛNҾu!(k?yf7KXT7AUr\L+ŲV'gm}_?-jȊY%g0TZu0V(0o4~!xzsu3ξ)ԍnATIH!!jXFѱAĴLSAYYpr¥ICC½ik)=UK~$t\U>M`j2j6gͤ>/=`kLo=
V	^n56"ZRczF$d6
MUa7Q*}	C0e-H"Ȉ&)ԠyJkmEؗ$XTa2L!Ţ-dIt/e˔DESdtKGc:SܗLSbY
V׮ր]D_EءJTn^Doǡ__,-ug"8" (A5	I/t*Ew*y1@oުUTeZf47]8b_}uM`o~9΋e:ZIL;:I<|FW/x#;tOJN(̘a`EꨒpM5A>Bn=e
 P}ߖS.uVXz+7ƂQT@m_rczé8sdI.H$`.ҚLGeԯt"4(350i׉t6o??VW'uT/xnDߪr*N.mpŔ<J{[xALul',Xul_%k'7jeaYgNqtΩUh.Uŝ"$L"5voWL0]&lb[))(5;~g͘,VW_߷.S\ƞUpфH?,
hpՎ],Gƣ1|@í|ZQr?9}|||G٧0/՞JU2qbRLaJa|f3CSLk˽FwKgϣ/9Jx›2
)|>jy.#O
Y/>Wqt{ي4op!'C#BА`pN:|<7xy{_.\+ߋťRYgG6nvMgnzQk0>XSDijdXYOTԸ>w*0NWЎ"}㙋R1!0
(wuuOa0	;l_z
/ԃhYOhvQx<2Cu8^
׀4|>PD
8	#
UmvEN[/mk
nmQB9YiS1>:qar8Ti:Xkmr>k!t
I‘)/{GVZ
vlj<5Q,`=aLi`Ta
[H[ L8qv|/ͧk)LʌCbb\d"*qX:krn_oog˯(/\,EN'3Ñe`TkʰBKTV#X<*SNoIzX6"*܄xo	=It=I5STidZai}^]k[
_RdmV1v~v/QΐRPPd\ia%%sKC[oH:
?7
lɋAރ(P8RҎ9/?$AUƎQ]ypzx9i&h3qTC4Q(nVnY&c^ʣڞz*P˩&K..˕1h6v43ͅQj7Uc@eA&σ8̂u@3X[ᴡ1F	)̌2_ʬ슎aMvJ	*i:)0pLE:t-SE믃>>Rٜ癀s.Sy5b{][&Q5)(Mx
PPfpeYPPiݭQY
&CO7D]HdmV|Jݾ_z;9H(ĨHyHJirNPZqIȷ=rJL29'5IP	ħJ:Më,*&5T<p~=&DmǚG_hkW-A
SyF&)qrjt2zO-Wk0,^a}ړxfSsc8@^|{^^M:;Y1gA4 lF]Rⶨj>Nƣm?ditqWᤝGQi<(LH+sOFٍ`h.(&jgs˄RXb_^ZlwE,ꌊ⋟vSTyuQJ
ưt#qIڢFJw\(Y0=IHXwQS7Cc)5%X`F!֚\h41(8yRfyV{1LjԶdڌV ,E]m^2X#[럛h;bx>"̈۵p6KQJL9i[ZSAx
**HqY8pv{~ Qj!:$@ѕ"i$	Lc)Vg螯JGk[/u)*G15զW=0u҇	Z?-?~S(-> [^w-sƓlQ4f9>/h%5;>1UZJ3%xH-nAj~jؙ}i0 FZn,;dqILXˌ؏ŁAKKy4	R'KJDp4!Y'3c*g8]n4Fǹ67dAtl|B	Ji%t2{;oMA_)"~pΤԉHo+7bbI˒r'Ad7ח=>O|4hDWCvUݽjEF=iixԡkR)
%RLv(DDkyQ6vF,[ZlG1g}"Sob%͘Pm0"l_M5,ͬDvZ%kqW:݋ފI{,4tJ.cL`}HNw4RN(#QOz;y8`1O}59OOe)/\DwʉĕPd-Jͯv럓vLWfZ]/V1MœauT#vyv`ܺK5#C Sh4^r3
+G55cihJ0yhƟ˹1m=Wю\;ڰ/JV;uHMr!讼-F5*
;aTey_#FՍϣSI1Ebn|CL>{TLrC}!fN!+de%1vEd74ЁU'5[b`*tچIqW;Fbw=м:iW)M$֝g1+8:~l`E5g/rX񳢵KՁ7r:6 Ck;~|(wv3?tnǰ䔒"cUIg].&3Xk0ΔJvi/Jһi{+U!QdvH$΃)`ge<ӻ4Zcx{W^TIqHmЉqy`ũƽvn-zo>xhX+2%C:%dX (KJEuk҂~Yrj'FI4|Jظ-Vd2rHj̛8c^иЬ^4c7nRl`
ô xoU34UaU1b+ioe!yt:٭~X{>_0sgS_=i.!/j۱Y{'Dyuu˺}|tzShWV{$ĈkxI&\(+NWmCe,kB:/@z=>$\3vzoqI߆QEqGn.0N9\>! ;0hL36}ahfMP+WcR!ᒖ;E@4Ub#Nj}ϨLҷVF3*@ʺWYx5˰LSwkb-$/[`1Vr^,܎c[Pz5ؒyđyoYr\23=Duü5al)25or[W~LbvMyF@кƹlGѝ5oQ6?(Y`ck1ETk3[n.$kdiI$	
._4N(Glr1ˋzv,qvāOAr>H
Ǝd5JT0:B
FNǻ]9I YVmG$$6>\U>HEu2.yh7*?*AXO=oMja8?zFؓͥ*uU䢃f+S&?\oTag~hlP*8Δ)T
lVo?;7JPK].y7nzQXV]aViT)YtR6{Ya}"QY
!3 0%<7q3e`{Ntft3*-c'{l4d:]lgqYTzo:clVÙ&-5'Y.,w.le?{۶Š_Y``<lڊuDN߷Z%Z"xB1/{}٘+$e"cɭEhcrU3ch+?UC! -ϖ14Cs Tn.9|?!xRd}]%MP)葽wk'<N#R˛|{tʽf/S6l^>QaK!NlXWXG?3P6L/Oܨ&]D
r`?'7]d?-3gֲP(Yeo~5HL_Jc*Mи)=.͠DXa%$q_쑖d
o?b[^mn00=hXw?T9݌ِ(	ߺT^'et
s6_hzHctܟ$~6{N*T04Hu[]
Q\r%ikAP9C#e#;.AQ)Ib/?w{n=OR&ͻM:+X
h@zOrB1J+.حsLϧ 9G֐
O

+uN}arzOaaUoީwT/
>?܆[
_.t'XA5/44^p-0)EU}'ezo.#766t`MG&_QJ%Tk@&$S[T+9~򰊓q	AJUt/Ra5G	L|9<"Eu23'{74~6G
vQTpr׃!"]S,imp@PX/Ω˛֋Ult.^_ĠhHRK

[!%4.8H&\^)[k:!|ۊǗ$#(n
dPN	~H
+"Wp?ʸ V94
Xv2R?Z8v
Nn2zhm<8@:KRV<j-+xfP``W4$
3Gܐ&Zy1>j͍7|z[i7瑩~l˖cMǙ'#XJ+,ßOGB~k4x}P65{AjލGu#ʟ9{')|(?䠃k鵸^
ck
獇^j^Hsᚘc0gN`r.DyRe:͠F`tؐJ__խ?0U]lmJq:0gvv(
k:$`Pl!$`
a%V̦Ѳ[M&qC~yHc1OY>%5_gLڐꥴSqw|#+*YGSɪPDD[%#J!po>Jպ}1N5$weQ6>^#n.mzH	32q#p}oF2vy0~{QfQB(i {x?j
ar])C\[+S):k4b/&i+ 
N8F׭[18ᑱ%X[Jhl+כ3m4Bot^Q}wtxkU)\ʦ#'OR
|,ËbGm_-}WYY%4Tʐ#(}lX>ϦYB[挛1RRj9fXV	d@ "exa_n*i	
ُ&F	R-}-քKj	}vb~r~w4%rNzG]Ϲa.Q|@H5/Ӏ D異2kTX]k㴫[{{Ҕt0WTA%4zeKl[K|<<1~yфler'PD,a2j5FORn1zg[P̼*#SR5*5F3EqDbĵ%Q|ߒS(ѭ%Ƒ;yBoIפĪ9ޕ^p
+qy<'M[wp5o]j*pT'm\hpC]p~XHHe,v&U!HBkpR|gY6:Fo2Z%~X	
#52pCc?_%&˗Sx1EUzLQ]1EXSd+!gbLik4c!}4QH*,C(&\gԱ
^YwgVveѯ9A&$W*(TBc:pִ{goJuLa-"0Ȧ`
D1OXFSdw+%-T
-Y,h#dwUl=NLJ QĔT/粂
chpi2	-9O8O-z)!W/ҮxÛߚwB:*(Ws`$*c(Ч!x8?{	^z|אV+&a^cmUj=S=5O:bg@TpyRUa&gy>_ǃ<=wiUNfT2T#EQ/XUM멃gl'yn![`ѫhkϧqjU-иĪSy\TIpɤ/Md;ĉడ΁`1d,H%4v\޺l1HYc0p}}_%<ҹ
Yc텅׳u&)5fq[85hZ*._DXM8oX8:]A
7BEH-,e."m9R0l(%pŷ4>T=ҬUɊg+ [aw$W(Xmg٫\#UWWO+( O/õl-<-0B̿{j,-mp{OnGO-sϗAHPiL6ƅ3Mșd,E㖣|G[/ۗ0D2!"$B2^ʵ&){N* KpKh;0p@<'}B%I!P>$VirJT
mRI
yz)veS
zW2eTRUCr__ mi7AG`=}oXcL,ȧz=+dkU	QA	QDX6[slVj8֬>6qGT˗[<, B-=tM2ت<&gLtM\cLC]|[dD9gNbTBcB#8;ƙrƙCZȦ}qEq5
ٚѽϴ֞
,kF+	\. Mu.}JCo,0EcH$4$6?pbsH>? 	.Aeb,ڦlGE3ޯ֣Bޟ4͈4
tnf1T=CWB+s*kQGߦqZ48(0!FcQy>	hnB4IJh.y)BԓZW՟9Msq@x[s9庥Cz<7|Mlf\1ܴpPԂݔA	}HـTRVX9rHָ=7pNd
5NO`M߂,\x1DVx4ڂI =(W#o|l$bbHxfpZy6!8:Wae2MN}bo~Ǜj~oDc^Ѓ EZ+|:OE?0x*j
l y_aݚPBJba044֫E*L;B&#뫵zP5SRneX=LhBpdg!0l_4ٮbhcvxF"?FmoFQߌw7F½ã5ֲՇޮ&p[N\? _\B/[pBfx9%\8p̕PN8,a
HC8NTp*)W::cS(GG^GcߔWл,V#}?wv455~ivDh&~5G)Y>|Y@H0;A 0St"S0˕"H(9_|(L(kz8B_}bWGאha-LaLޒqEʰDrJh;94t>3"c>C
&vh4-ee*m,ۯTH<rZ0GS6"a8#c#αeGsCh.wYaL
%V3s
gUS9#DBPϳ/>M!Io+dqXF,}8|2`f"d?@
8ى}E/[h/@clv<=Ac\l
Y3\(aXek*"J⒐($X"
hzAiYG((70T J_n"5bL\
/9YϬ4A$"ߜB0֙:JЖݦ'
 ($̡Zb(T'()Iz*社Iíva9{J	,MM/1/%k1=d>!˅RZIgy`b	xL( j)%4Vx98~lPTq
X}bHe[RB*-Y[fV*%vfߖXLrvmp"nG6T9M$_|O.1_qq#8%|
@87'IHG1as㜍18L"ViIqYĘjS?tp]oGW}~UWWJr iɖY_5MCTf4=AW]/eҁR2V/ea1_S\h9JG$k:%2@ɻ!+tl=K}eN^Y#btR]'zKXBc,+C6z͗k-L$J/^u<,͓|ޭA{'G+ 04EXDd(*٢.jhlu4m"oq7Fu$m9R`Կ/?Z2{8ø_)ߓrI/2Y
FaQI/$O!~ϯ/:'RC"ƬR6(!j}*_,޷1s4lfx^k	Z	a{u9X 3Mz7/)6ix4Kg-ͱ96yCxhzHXb&\v]&uhe{]&`o3@w$M#?^-IixywVڔykABi̪lwFhglw6ˎ)?gǗmUtS'#Zhu2ɈnL/UdKPMTkbCVk1|U5+XX!UPXKoݦ`ϔcKs0P5Nov?H.Ɯ}|sqAGCJ48a(J"ī7!Z)ʖ]
1h"YLibjY{T
+ב&s8[0tZTY@!WNUXy"ֳ*3+RFq7Fl;Uȸx/F$B=q[qyޗ'r"*T8};j%
THY(	ccJazx;ǹ쒳1OT>ws~aOg1r˶OV\jU~z7Jrw/tzNǨ0%c^?װY۷T^JtOzy}*7[L~կv^VZj*)"FD#>bX){_MPbJcioz<2eV60|NcDoE_rFۓ1Uka
"r$Ȣd
8S+1*U,(i8K+=h,_aoXْIjaJM	'g =֓o/Xqi%Vc2'U%
ibc@9hޏg?k9
~lsOj>[ԋٖ,e(YZ?_FKeKKh*Yy3شDJ{RC%-W:d9Fvo,o;Gܔ
*>7Ye|!%ZVf9c&XWq+ќ@vQ!\
T!K$1n粃d@t{NTMeeȂuAKE1F\c+u {'%4*|,?HGJh3^6`q$*usu>fr|4VymlD4ķ,Dty@C&XXWp-ZlЁ[Ju 𒐱'DGBV+JKB1uJG!ONײc	Ҳc<{c(z_>@b&S7'>n+hK\k7Ґ[Ŝr}O,6i `ِeY.ҤS?4٧	O+`ډexrCf\e1XO]E#K"іX]XEJ͆469"䋳*]+#FY^^n>WCȎشg8u_
5g7lVTO~0gv5R3SryD4 {4ER̈́m$%?i2q-i0V1W_)&jsN)owo^Sx҆lCv
Y'pKF+f,;JAKw|osl%(H䈅<gqSw,c	
%alABuiU;d:bfxNQCNU":bXz}L{"׈/Ojrr$}w1{G=uZc̯FE3yBeSݾ5MLܸTߜo.r7[~s^"3bC]{i*{Off\63pѝحa ۄ:HF(nVCkjN@	O5H.656c	K|oh
L>M#[Z/ɡa؝Fk7&w{`%X"X'!thx媣a*#vK=[JO#fO?OxœZWX46AT"]46gr3FK{>{tv^;
+K+J)o[F(̎žg:QGZ#F;r$SMVFc5)Rut^ڪԘ1BRrf|R\\3jQ3[6{/ҼbT{&k6|za,'
A_\>9GJ?hekU&.dYC)Fٿ_*>~Hj(MY([,**>->b`Y]e+[q\b-zv|ұ4ZrmED:cOoF56QA;/ǎ6K/<*9?Y͉;Vwxִ[V.5T2Ue{JF%˶T
k-[^|r-!hnJfU8^)4d7dڊE*Yh0VTR!D7K<+XeYTj3D`P8K$MWƢΐ՗]OmhE-YqpO9ѿGOm'h=9D!=9tlQ*wUiY	W7y?޾Ⱦ!b̩*ܧ7.:nVwZU8<'NE5?삕C䧶Iuʽ`]-׬x~Ϊ	Rg?{~}0^Wa[kfOV7N'_4ypDc1sS[G3WMGvnW	LM/JK_a!
oZayc!M򴦪6`a;N#,p(RN
o4?dYVnI%hݩ6J>x.*#\<D qOS{%>Fy'8sa9b&B8[4s#a59ufQn>n9ӐwuqyW"QH
0}]b_0ig$d/Չvvug U9GQytsģc?|r9'PS@o3:ԭ}r&Q-&>sν+J{@{MOJv]F`QĮč!'񺄠.%Npx
((GM\Y/]b218{ul;0333BtVHTMx()q$ߧՇwkMR2"VĞ͗Sl>?kT(YGHē=
ԤJ)Ķ.{UGpjAlj3B0
M6cO!c.\b#,{ͽ,k?bӲ;Gԍ9FO5cw,FT6*ŪU?K>#B-v\wg
	rUc9..+q.\ӈ+Q\B˚*ralkvqlAӱaϋ'Blj(s6?e(al(S!i?s7ɻwu"8
ߧWnL:@Q3Fs8^ppPq>wcv	щn.VD@[3~=TT/GN`LQHjfhw}7K;ۥ	`%o4JSOxrh\O%*QLUw&+H%1Prlmٰ5Ш)T^BQ莰89Lx45_zv]5HI}a,{.PIu4R55aQG3Vz,cY{Yz܋?M	7Wr6]eNAh}fw>h2:~HKoZ\oe6!٥
:P~*AT=YO8u;	,%!eׇ,jR"gq
Wj2Q䇄`doiRz)#S$vN^(E:K~MDq1޶,zEqŠAɥZ.9$?:(9{ꗂ}cuo+C>t`ϱ#gjxYz7AӮ(^+w#|+#̕F[P[H6+#@rA0:	m՗,8ԬkёLvG^h>2=Ce"Rq&TS;3]"H/R-#
)W[c0a$RR/Mʪ.B((c!.R9=;/yb%˴esC`QS]Q#2L{NƷc:uxJ 9+-J>y;&+LI#kr;n%7aMnKWEǖЈŠ{A6v劳VVr7chK-iԬ0yTf>hoV#lH7v//~چu_^1~?wg.}O}yryyvn+~e_/듖N/OKorb_ƭo[t~B~a{fkuk_oQHo7Q6F6)5SѠ 5*K/hǗ7.חۮkNA{ylF=:dZl؏6v׍p߼ڒZm{^So(ibɤܶ dMsfWNsfMtV*CԈC׷x?\m?߮viCM@o/(xyqءt(>h' 2RUE2LE۫Fyk#OVS=K۩~A4
ѷdNdn]a@܅̹5DT686G=WQB
H8W2p9Sj۰6W~Fv~^\|*Âao-op$n^*kcbŐQsLa(dmN+55ΣQkYMԆ${$=б$
>G7b#m|UŒfԇ0>Y*ؙ~RVͫ`"×+C@b
:VEㅢ'>;lf#ݫ
&\@,=X\YKoVPѹ@-HpFyykN0`"4g}]I&\|'~'ߕg[בbעv6k̖kLNo]/5BYs.z֯xI_o+Q U(Zye^y*1ZP%U6U17:ϖXǕPQ:U7ʞρ
ƪ4.h1PJewq0~2&o#&rO
ɕчM 'AUƒ%CN3>űAW;ԍCځn@HAcs#ЄcЄa*Aa
s0XpOxFm'Oڑ ZzWߝWqdvo¸,ѡh3RF̋d,nv>nmg[B/aH	|lnSҋM͎!Bu	Be-wn`IO7V ηźb㽔n4eN
-
%Yu:-%&Fvj/ʺ!@\,dK|aќtb9xy^bB0ͮ7\oUOCs(+e*e]qlN'ΕCAYKs{Rq(ͱJ6(((CҿRgc{ET_P#"QE,sd*ԒCc1vclh]lAoQbsVY]U\Nڠ&9bOZq
rG{a*7_ym3Y-_I:(=:zC1;5P	2voEhqHBS-u{X N-e9XquAG^ضk퐟d?:A@DG&rۏyv.|=ykǶREcE4~y+O6I曩y^UNBwP{vim|,t;4jtaAsQy{CkAxLK_`\뷯ZWM̺emk 5KAns@Z@NkeI*|r_}~plNKȉ=i"F3rn0@
_!d>GdRv&s K$'-RH!LZx:W)58$eE
qy"Y/,ͧTW={3YH6?ٻvNsw~Q^oOO>~:ًσ7F?u'3#oOF/k/~ә/}|h\"kW~ P35_WAu:3r7I&_>gNgIzx`Z|Cx6pC)-f
rt˭x"?\qKq]l̅άtk>vsG:=u33q!RbĀ.Јa+T`Jf­c5YJ1WBY(ώia@^Q_5|xk{cFf?Kqg?(%rss'41[p0C&]$8Bd8s+&r.'#{AMQcJ%Ó0	COè%Zlڧ1f[3рYG0f?.>]v€pff4Ab'9ɻsjh޹ymX@a|Qz=4Sj+.l߀QԘ?>VOJHziքM	Q>XL_ĸvxvb[7ko^5y26Cɧ%яΗO2ق5ws?o~!3NJY=ٯA6:)

$uZ:t4ټѐPLً@e==n {Vg]`j1}[԰;QS'U-guJ[4)55J^bM]ұq&Bѕ~rkC^ȮiD(N=CKl)id1Rak1D_F2F[%.2fd̉G$Lڨј/O;!kB1;\z"~d3
<\va
WM#
2\mKp[+	r[g5h_TtZɹCXrž6t5VGv8h*\-5t\a0-h^)7r97A;P
tFb}y2Tuƨa\(ǰܯ7Xx&*(;r70#￝yg 
.Ksf'o,B54]65̡=jx6Ci-"A!m??=$k_ZT~[ŐZ3_'ys	m=e'u⻂xtL7S,#Ő٘J|!rXiV76C]BwZ[ƃٱqUO>M/z ڬ;-+hP{ry^Ɓ7{yZӢVh]LZL[#p=,avqM!w;ؿm_{LgCu#{%#P/3iv?@і⧳_~޽9[]aZO<^߽`;eF6""׋tt`3	br̹s%w	3&Xd>C5Mwjكfu\lpm1c)#}>{L#y@}iC"׊4Ǻħ:i~
rlXo0a1<@O-DH 
1GRkS!*`hěy@@7~}LO4gEGCދ[<'M~-r9iXkDJ'pQ\=QjJAi\>
|EkJBԷNXJ:җ6Og)yLi۵
>%.$m)L%R
aUCg.GC[_}~7fKHj^$9NSǑaOi-o(.NU~|| GJq*{T7ϛY͆N\rNN8KBVkeGRRd&1hmF#&N+zm#K,$ʬsSz0ݷ,yXT>)B&]750ЎaJ?PMDoDR;4vY%Q)Z7Ԙ'DC]ߒKw4p,Ŵqq'C^f u
`3
a9]6Db)jKR32*6Z2hjyG2Ff4)Q
y\#q&|3v=^kDLbkl(ƝKWŸ%wp-XQxpnݶ܍}0}?ͤZu34R la
$t/\VG赗Erk	|њdwͼD4ɩōl\qv,'p)FM3HER 7aR0`梈~[Y>у9mc@wEm\y:(oI6P@ג>.`
K6dH}vs$){7nVL{P[x3݉vuvųNTBTJ0*5V-\=Oؑ@ؼrj/	!kM%#IBRg~a$V(ym2Z6U[6D3}~":wj%zCw6O<nKSiox%ț}LG 6,~y+$O6T@l7UEc\1`}[e+|BBQ9a2:sQA|tg-]Zpe^]HȣӠ¬:4HU2Yji{McKyXP"y1\IleD^ ̺iPi,qV2CNJiam2r
%|dJ{c@V]$OH2)W7e-y49fPEljyV^Wj-s3pryx8ޟ=?Wds0H\JG4<FkbB ?U Y3Cb_eLu.j#8}ǝksA(HgMj#NyK|ӼrHxtIb1LXſM/ӓO޼w`T0_WgK`n^nqmFX*]vq+PΑ2z[
ї웖nGcqKȥͳ*dc'$@a[HM(Ąw^y2{
AUp[(ƸX$>(ʳk ?G>}f(p&
fR+T\s]N9_z)EZ3IC-Qo3c㌿>*REfSs@6bw5UR
˒HH$2Jdڅ,sRޘRit
QXզٿh1'XԜ1YRmLEbe2\ւd}\Lsᥥ=W
%f-F@fav+	C%@	:\8ʈa9~#yPvjs=q1&	x]3j
yC,
S$L/@ݵm'}īHTs#@WD9^U[ط23qxa+½$@mb=?<#fް9J73{Ю/8lV
!׺h3f-Y@;l]3'!Z4ޓ:#pب=qd=$V,o9-}kU>`&YMQ$Y[k䈥)SXJB59NHr+yOSlM>S4RF%IaDB
|kL|q>ٱpB2SrSŦ*8BjnƲ0ORh"Zה`BɫNf]SC1;K&884BDF;D^K/|6֚"RʳEevB4?!"l#DKI
9*"FH3gę'z1$DmI_un6MpEtdp]jTFDnٷNhqV$L.Z]zD'r&IM'X}|Mdr)2bL{[QvV+5te-5d$EdX2Z/
8wQ.Y&Xl
c1nRi1Ufron=w[ŧvMΛϯz{K:e|Pݑ[w}sŝIiV%c?-619rH$B3'	>x	D&DΔC66*Lvuf>G{M;VrE9{V3-Jn=wbv0@9)ЉQKIhL&%Q%\IW.y6DH/Sn!DDyzisM`6R͢sVz6L'x'XnI4NM;b[Jw*Mh-ɵEzWF;AJ5t1К.2h}"Z^ee;dͪzREJ®d;ISX۽i$m5
|MǺ֒p^o{FңwNYlEw+ơg*.Yݦ|ڻGJ
U\(,-!#ց۬^;@Z؈]䓌y$cԜg
Ѻt2f`yN4
un4]'Nj>n?ܽ?;SW(Qp#N|+.=Ci.(x
IT!s	$DpJRw5:E	Z8
O;m,Rl$zr|lKGjn no~_B!uj`]tAP`"+BVύ>D7$֌rXUQM*gOsoWN'(dSxOTbF" Ir'()*66F]v9Pk\AW(+zEӣ[#ޒG
L{s&p<*NE\ XV
X]r6Bn`,kO\bLPT1$:,?l3++x3Ta`TaYLf*,zR o,[2/`9}V}q
FQg+:!	8Qx0VdbQ*G.Ql..3|VThWG%S|q+$NCy>'Sw'_
-[;a0^ƈ|,ktQzRUܒ\<
.,ezp-Sz3IDݻis%NrΣ}yp^Kʣ:	uJHivR9vΣys|c6n׃@{7.MV.`Z7_Kqp(D|>\^f|,/ܟiO{"Lt[qpX|O_j-xmJג_n(SCwDGGE
*|	2dT@[#"Q("oI)+#*q))Ĕ}T) r>=UOj꺕d礱;qq[/lL#KTB
bh|D2A`a=֮mEdEږP>&*V
MZbӔvۀci4kPd/v=Yk+n_3"x?!Qy]ھ	ǝTQiAj] AQDmA8l.IV+IG-8jpRj!I
GI1;)fk@.N?qK1tgj[$SeEuFyKЩ@+Q
Hw,xNhUa^s)r%%d.Q9Wbw8eD+>_$i$SKSH!גjB*T"Eƙߙ&VAfN|d&֚|+bBtArsDT\G
˨E \tTPcK{_|`2%
3袥MAe!-elPUM(i#s>z'y0,ipa9z fiw^u(]GC6>Cޒnbm4	܌y6L(`rnN+-s-3;	޲rȼІ`dzxBkҒڗ^<ׇޭ#v5rw~;OXϽ1gz	䗔DV:40^[4I{mQ@,*܌y&P'
kU%ښX$+tJbp*Y}XXf&wC*$3+IL훵oѽ?/a"~"%irw/ڞʊIQoO/f0)F)wusۄu߲1e5y"-<3;}կW+k ;;x	ٳp>7c
l8-z캶	ZfܫLߠE@ݧE1GZV.]ݮ]k]KrH	Z-fI=fh^RL
Aj3̔rn0Ǥ1}RƔg_\+U8;<拿sCL|ogu݄JL±<Y9~Ng;H7ĈUvHEC5X,?GkIB0GkPLWuRLzњAZUϫK-Y	',їsNB1E~N٧@?o\6a!OM3:oݐo]۟ٽytW[zL'n.p\[u_b27NWO8<1Bޛ/[p!]q6E_5Lc3B%x"ׯ?qW/߫˔nwş]>>\13YBKG99nrhp(z'nLYhUݯ5ڨ
<
sEI|EMbtQ/x}#/[%pBD`rJ^ջW" W%;jIu[Q>~\Տ?<Ტ67WOqG	#B9c_kH2yr-:$4a~ڑ9d7)킖nܾ_}qvHe;j@>^I~x8wy~y0_&{#}^XNN/o.J+ܷ7lzu5Ϧ'ty"K	#	bK7J{@I6~B`O'ͻ (	:Dz
Ӣ?ՙ<9
TANa
MPS*
TmziПW&o *Ãۛ2_;*eUDja7dh;=$rпIB?/>c
3-Ŕq?gmz'$iQ\,—.|_&z"ZIloOO+ކrHVfhF5fX=l1Qr"\
6	J~2
9e93N̰7mOA
ŃWXgM
n~?J(Ď^ebʭvƹV-7	H]k^'	@zηifb1Ow~rA9:@-m?/c[Lk	Ҩ5?5M
W͂Z.iKUʢ`u-ʅx	'sFȻ:Cf!ID@l
Y"S+Jy+pTgZ,$]9zA]=8O4E~W.^m'ү8Wyڠ(/H;q?s(ˢ0dB|Ұc>:$<9&񜫉LB6ͤ 2Y|ʅ|2lzҭoo޼\1љ\!֫΢u\7{kcEY05|{=>_t/q6ĭq~0hTuܾO}"RjdBeW1˓j4r)3mQ}BWе\0dqv3_}Wg5t0TO~Һ|=^=*~,~,dσ3.ؘ`KYac՜p; ҳo]4GE$O{L!?xOϸEϩ3t큓T+ڕk(CU\${((UVe^3#
˨L1-!FBUNL$7qO-EsL+˕"]8DjVKI,VL/ʾ):"^Rizp_
{6
#1xxDf[C<ؔ6] =&o+p~X2}.ˍT`}cPIV땼ߏ&q/O\
W3x~4L'N(}fL3mSP0EE_L_v\)i(҈\JCJ"T!"uRqR+,ĖLl{&LB09uap}c:bj$}9;?^>nmf41jQ=JgM#IYQH؎9?sM_"G +E_\99{ϭĊ!hf릉ֻֽnL݁vHF)*I Ă[$ՄQ7zOgQ
Wyzp*ͮ7Ef"9XXCŲ!^D](bQj) )x9'5YëT
'mn_:|]6}@
ڝiO\|^x3yW"Ë8exvp}0 6b%]~;NXpBb"|:C#YH߷ܿ(ڜ,US$vތ>o+*7%W[*hV( ـ	\gX|!}k[U}{gkrVkn@e~B!٪Mp,q:fFlT726}77ϗ`e'VnCAܿ=jTD66Y/Кk5
yBYh~/~g_~?E~ۘƈWww©}rrw,{.ėq:JRqi,94 =q-PkN(LNƹ)TqUІO]\ߏs~nONs{Aͳ4gζDQR,)=NX£E.sJ㎳+V8Jcˬu\*E
s!4nb?^EcJ4
6M@{	Jj*Er/Ҡ`@r
(u&傛
}i`b)S SȞ(,82uEG"r[9GHDc<•J?KJqU">xD0ցF=3(n>Q>@)<ϩuA`Vz\ɨztFQg*j
T/He7!@ԐYNJݪ!!مn$=b$4y(yR(z2Gra)Q'i|aF#dZ_܏gQӨ+%2lb>=r㭓 1=h?\{52it"7\MSPjA~~ri7DG٧swZ+Sl:x)@\BW8 |w$ZwQW?}'=K4FPxzl񪇧v-»+mje?w7Q6LR*
7	%j(m$PG~l#II3(Z..ouxw@,Ťa}dC߈Mo;f*7C[3(w1c"5ٟfiNDwh#GB^C3HuIERT5[
NCh%a(0EI-TNpXP Aֽtop|FIdlUZ	KQwYYY 1y@lmDZXIZo:cb-l29fÝ#h@'8HTVs@3˜!S9q⿠
dX!٨?e$ܭbFlcUx\X#.9āRWwk gqt
A.hdkذbI '+rwyB5$G&@6&[5>5W}%G
FqR|jrh,+dhA~FF4h0νѶ\߬pP0/lQrMaB:w\\NsQ&+e{+	[եj*$C]
F3`]S7f]mр
plë$Y8;
 wA)ޣ/sR;,]g
[:*-9SRw\H7ljt\@"Q	Nf1DWْw,5K47;j.	чk;DC!EKrtƃFe֥Dpx
Ϲu,PX5ڤ<}!m#sݓ9}cv
U$mc+?|(࠸9hX.Ȓ+,%YZ$B֢(r7ݙ	mp9^gd2{¨PBVd|Ulcű{y89wӦV~d?Bhvgn믿/`%6e"S83gω2YBvKKS|fJ[
.C%-d
=w0ifVoк7їiϯkptj߱8nyҏfx@&y)^Cuh9\5y֌< ΧG>q<='BN3Ƥ*~ ˦k*8^9=]P] 	^,qDj2I.0)x<$Ily1B
u67Ͼ[^
LOnqfKvvIL!-$M8}
6WIzac4§DYPs]y}~ӻ?bT,1+s[r%0>nr`d˝n8U-?NUʏ+%jΏ++Zӟ˃՘c@W!
U-eq1usJ(wY3*WPik:RÆK@!akU}QhU_":RuQZltÜ%))R@;*`v}G
h@;\?2+@
D6
!_c|R~TP"UOceM3L8!Ħ\T{$tf4|fuI(/gH I%)-kbe0+HgccWĥsCҎ;,3,т	",CYFQD#E+#q c4
#W*K*CYE^_0sҩšOy3/4$,8Xzxbyw=#q/Np~isO eڄglXJc.e\gY,똧oϲnp'^/3;<!i	jz,PZZgiPS(@#ȩAy/vذHX> ㇫RzMzvX?qu2FS&JjG[Pb8AkAc^
wc_p4Al~<)%KMR@!^..㳼3ĜFJ̐%^HJMt=WƘ,Zl\[GQhEݎa%ϙPfL+SW8,:ϑoBj't 2-L+Az$d	˘8Y~+i+}+/⍡k=51լ܇'wŀ'rN	()J-ͭLCOTɡܠw4[=--1x;N*!Ntv+AU|(\7hnp€v؝Fٵ8ݙl8%7%_cL"{m=53QS{'ܤ9.	[q(lPWp(E+6b8<S(PcJ&y-C9fByq3@`@9ΕXB+ThVn]rK"#HJ/5q*¦0mL
@5&&D;w:UI*qWful]3ٓ
DbuWlɪ>b/brNRJljbL<xT@V6*(^JlgsR]oڙ|c>!#LK0LaE=όjQD2~|F݆RMl9@3DKIr^zE\.vzf̤]=1bmͩyspF⅞"O%R2iJ(?bAS1`(Iͽ⩓I) ͘W0<&'ӍeŌL#ɳ1|H{'fQS]p$Im&(͈A9!
FNAFP
$T+D S#ѩup/3~FKY^?NnQEI=/C*V̌6m5ɾ|о{3Em
G-qK"_\tͨ諾٫;Z,ٻ/_,Qq1.678awfO2$SܥRR*a-TGlˍs&/ ot|8J/?)~M
a	z4]JJuxӟ|&|G7xr䳻q0MYݗ@l@ۜ@Wnçżfu@ke V넃%,*,DzbhqVeg}w}Č3`޼0OAq3f?x13)IQja|-%g4?	6͆z/Z/&g_2>_?VXG0'~oizTFrMIb^L`ϴG}0<az싉7psr9
&_,d*>^`f@B+Z?=LLr tdte~w_J/_o``W',&=?|z+Imqkq}2|љ
^6ڙ?drm'p}ۯD_F˷p׃{
Ì#؊|2wF>${ӻM߽sbvenKGn(|V;g"~t3P`owi~ihxkMן-`ۻs-K`l/vh?Ũ&A
i:nf~Qx_g|̯[ť7s<?
7nd]0͕7з~~dxz<i·֤3ATg7ooG{0?{WܶF_iO38mO:MΙIDTI[.HY%*E6ibSX<,w_~z5棺><#P~?ᯣ;	K/Xgm<F&f\~9n%4oQ}NF7c/e|<~>Mp+X濿yO`^ɍu6a 1xWoF?kqv9\L_,boF6߀yκAF}drYmn6RD11e"(HZs无6uZL|aEXBrڥDnPwg@3}B<l2unq,햐2<ۨ<ͮ

Ć'%+rɖK\2KNE%[.rJ\r7pIlmJ1`."G1R8!p*B\2Kfʳ!b=)$RMӸ[>S4`<SoZ摎WkUX!g~bư$fx:p<
'("0)tj|z{bmuVhv55%0ے*t͏8xم	,kxzdVr{%_ͥNIRt`c.Sbv@A|XtE?نh:|HJ82:*M\$ҠƜ&P;AA#
WZoF-$
@0eJtTǬ`v1/5Wi?F4G^۷ow=iK%,-{ۭ8ljcB}Fw+s)Kd hw`{#ۤL;JzQ1)ѩ/K"X&1솴nQJHlO{
+hҡp|0h0J7<1>@cSےx2S/龲G^*pE)<^<\=^7G$L9)̑7Ҟβ5Q|@L{!
	&'rwد
mCbO($O6mCbې6$vcH5FÆxX-U	K`c1L'G+IIi[rT9eY{
x㏕$V-y=_}Im[pj$1ӊpوi#(kb»_й!1Ӝ%QU`D 0F#*>'Q=PY|A_#UDE>E"a-o)qpr˷\~/
l>I:֑KXtk#2I!*k0+&9m\ӝ%QhX)ZND
(`ɉz\$>8|!g$D%I|}kպV'Z&uZUZm6-܅Ka|5md%[.Y%SW:T\!K#V82LKz4M̼E%	R12)dP
x.$P+9O?^:cn%olOw՞?}͛8Y7e4fhl",bƚJN_i'"QQ5YcqIjPD]'<
gg!*(࿾2خ0öC!jQNVbӢK_f|"|?w_Y"{w2_ꮡsPJH\Ko
eOv3	"""oO;T?/9`ZGٟBgٟwf
z؟^g4!;ַ9&ĿY۝zDB	d3K_y'65łv"A-dlOۓ%1s٬sF-g>JR3P`+/L0Jo9L`Zc'J]Q\rra×+j}uMlw7-~௻s>ܥ0w3]3ͿL$8XǹQV?i%.*+!b525)y鼻P4:{R%9c~TŁeS?q&3߸FQ
lp 1(OkyqYiR('VaFiaV	0[Q
\"_h4N¬u&=*1a!R8>@¿
4l$ccgĤd
b 
&7f2A6>
`127L#>}EH]8;Y	Ùl~8.襞nNg>{gd;gEU.;YۉƷspgl*6]k7_pLP_v0ˏGA]3ym0Ra]#U~X#(1(zG`DBVx#W;J$9II\C}F+IVN֚Z1a5,5ZU!TC'TN2*SA$[Y$#Lh4NB=TCB~AP+
6hth:p5N%mRʥq=ષᰡtY@&jCMV"8>lp.RJzYwW"-(hSyѹ9iܡ]-Q C;3
kp}ECC}L&D¢WU.zO
	Փm]8}_胃y>`Wm)L,wt_x@'5o,% 6SZHQSM>`&i
jtdxϥ]w+=5TPZ6s]Z6Ͷ+ëGWl+w,dE۽+tBۿtHWԣi&>(1D
wca-{ҢeDbgDhn+Z?Ѯkig6É|0*
8e-x+_n6FOYT{*`<ֿ?mX͟*9d$y	>,Vl)
-d0^*ݤf]rWl3E
DőG1b 

yX&J_Y,yX1|8VM73
xH6?;ao}ĈνZb\&U[\fjgqfbqؙނYr>:
m@_$KrxuOfO:?.>ep{7]'oᇾnZh)NFncv4.MحjĉAÄ;hǝpptjheo*8=
H\S[êth iJ	7HMoĬCXލ!JÈ2Zx{Oz:{7H{08`;l;2`ݏmf~]y7o31P`UGN	ۍ
`i,V_LO|j£*I,wڝ0tÀ|$&
#]..}os/S~jCk1߯x*BE7Sb2a|KqT+V(`8M@E
breΛfP\8.T0.X:blNl"`_Zt!j汫6L%5^,,TLc)́$$S(E5Fj*
bh+x>FceDpFb	D4A@}(Lr![%tҀ!ۦFYGἙ%/2ڒW_3jQ'W8yU	Vf;V!rS-9z/?u[dA2H+
ŒFq:ʼnkF؂o̠TI<I!0JybZxX.Vd!k`t-PQ}t2]udYXM&8reغy?Kt&{2p4/ٽ\ԗuY)Zu7nGn]E~:;!,zvi;/^~LߙƬ̵̪'W%7^YʼRCUNB[%S]{3c^3V8^dƵ&` D>M>&
\xy6aMnBCS3	N9gI@/		b~D8.a[K
G8W*;az!j
f*&kh@1H4#'P-Y14B3DIOcQ
0cP9JYRwX(Aq9E(Q)dbRJҗbqԢKq#I^;T}X=cp1OF̣uYR%Q$*7V82222eOP+"NgLPi)08ߡ5=;rտVo1xK
o#]M6eRj}R&+|EjDiiޚudo-
wIĤZ1?/ۯЖhXMAftY!G§iw@+@~/|q}m=C7]w}h^PƲxfR}Cc֠vQ)QkP/_3 hdzӰԸ#h@|6k?4J\@Wh~w+~blyI0q*8}BqL{(prz/ys.wp~=^w5pR\)c>).W[QHw9W1OH'^P8Wp.dsbʬHɍ\xk|5&Qv/݌oPŲ܇/i$o$Sך8-.ڷ2>zHiBdj~w#S%WZ\ǧj{056&WR7+X_.jwn)4v_9,͇;Xp9a/UPjdAIɉ`О	:X'	*moZZ0*`{\9L) ,?w%2/>[AAyf݄knj$%	2e{P*y!-x~[i$eAUk@Ctfplj篙zA`
QX;SLцlj,f9/X&
[)C$Cp{*6tűjHh܊T0I0[-etmTY-2q6vaۑ:qXS4\[G['&icy7Kw?.%$O38o+7P߾lQ`
`k.8B
6n'x5Z\pluO	sIS)ᘨFMHЧ l!'??E$g1WGJ+Ww7OZp6;<I
bi
ºKv^N"BƖ JAJke(Pgih˥x9yCO}qe+>X4R6!HuϔL.Eq*FєJQ{M`?77/]na鬎L}87Q{ !__Mvrͮ*d_[;xX<^]^P$+岆 (ٛv0WX_@CgD7[Q`dԢ\*vNy+Tr*vm8ktdP̞LI)7],|7BόJleʗ Y9TYHY΋NPr-|%P.QZֹ+jN
G#q(J!ýM ;\NʧXjkS`SԕC6 )=j)Q1UEt7_B k~KW#TN
Aj[{<%L	P(!o"^*FXuj6+0H`sX7-2VI;lH[ٌXAT+B6~	$ !Mq@ /v,jRFG/
$ʙVFX;;AY\~Er0>Fֹj ӌ9
d,2$RW/S^UV(w9%Wys5bV&0{y
M1nʛ~?m,+h5|3D^v5y]xz0h)WW1*c~gPA3^W/ᣡ~>IB;"d|PH%RVI͏67y
>B=7Y9yHZ>Sh񜏞 *5
LBH*DҕIHN7<TkeL8E*S$실ڧkS$j"hE,=ONhP:1cUψ<4Q-D{wv%R1jz[9櫕j{
{h]t12Td&
;kwf(nxǘY1`4g3~TQ=Uej87nU5`_x~3iQBF&]}Ph{N"Bܲkնx|BHp6 BM];f9zpO#fUw=֠d[1,䎪MTZv
 zqS2!XYVxO2XQx>x1/m4p(ɀrlnM%竕ױLjW}L
d(pډZIt0Aj2
u+JW~d%_'ۓ^Y"eXs1hˡtUۼIo
)Mnk|Le-QV*$o+1Z_9R+*}p9’*y+-]]D
c^@k)_&Ƕ\Nu9fs8X횻:7q6<|"SЛ{dBH:P
|zRQ9u#+U(*ѵPGkqp@Cq2sO9ȴ&UN]\>y$
˹cxk$FTWY7r9zDŽUg5Ì"K՝YoG-#Azcq*ѮiFp@%-3:o#RH8.cBb
UZZ؈i~R"zy	FfB?Nv$O|u.+D2ƘTalJIA$f1-cP%F#1WqsYMT
!R,ERU{RHWT1`0pj3̻Bj),u6yΜ.$Ry)‚Id0K6PG;vb"\d#e~EZ,TYXbLrpvB;k-X@!$XbUe;0ܹ)B6p		W$wӬd*ltYS̵5~b;:-1"Ddž&"/R3M
}L!csbrj^HqRe6BPWg蛰UJw
̻3ˀ7:+1HPc,Xg!ϬXaO+|.5ɕx/܁0X*ԫPӎ"EhTf"atL[bP"Wg-!N
V&aS-β$2VNj!	(geiP\FYbrV+ω31p619fN	FCX	Qid%x'陽;RLf	Z3MQ¶%Q#(*1jy0	58ݴDH8tdN	HivBz1n1u$L9xX͍Y8xDCE^VT|n<7u4fjQdS&#Eǽ._-C򎉏Cz't@MOLXH)p1;dh
G\D)̈$3Jr/ǔ6V].
p.aK,
c)KVB[Ljɹ[gVT_'oCfve?۲qgw`HC2dA9eg]fpB))H6WfᶘAU}knjѥ@2jho	uXAk;L_Q)L
	ۢ*<#Tv^']]i,*e"1-|Wl$A])fovTq(n:v$J
J(iϣY_(Py'<&.IHދGwB:qQV'OhxЎKLQ-N%tȎiibD*yz;*)8=]KvO'IWTDc#3_ltA	9cc(~&c5\ `N\}2Ф4^epp:ޯ͊Hu?׵y,`iKjiV0,;ļH"d`!5sPgbXrsT(cv@ͻRi`y#-
)mmsF7S|^ǫy(ϞؽYJ8TBF^IE$
 jHl
GUG2rB0Y J4a=tR0ʼn~Z+5$$D+cmAqD1F4CP3/`#6Q"B+2`js&(FC#lQ.q4i^0-k"5Ɣ+Jvr(H0 ۈ0b.^(E&nu1o܀A\g-Sgto"TR
tf3+ZYϞ`I9ȗ`V*(@EM([=.'Pr<.`;8;XV$-W3ǓvwTU=
_JS''kjGG)Nۤf`%%>u~.d̉j%GNNCd3W.=	p]UQ6Bmq,~^}ul{6}U
O )w({w-TGtp%M\6s]Q;Kxo:@jUu
5l]_6RŽO(Y y>$ZNt"-L0լTՓHXleH;69ԵBb<{9p0-AT)ܬ ]^Pb[eP"#)&gBr`:%Aq]'2TE4Fp>(+b&Vm[A{~Oc;|.@(J%Ք!)B[QX؆2@!AP[0fTۆom3A߮65BCC%BbR#"`#e|/
["KBZGm!ftZagvVD[j1vQ0t2d4 F&WNt
bS!Y8e}ͬ6|^qZd*Jv_&8"Z:6$4<0`놑
"fJVM0|6lUԽ0P߆QèL+e#2i'}X1)r%9RiISèL⾬"ϋkaivB9ꂅk뎗|Cq뛑Y&?KSRf)VU(a5?,P֜WS#2N[R.V;qmi1m(Ze2ZQXf!YShl֣hzLZ\pjb6-;{^N*F}i	Ay6Y`x@_RDzJz)cIɬƏK?.֛l>Ŋ[l	P l
6b܌Xnq-rE4[Хe5>xWC TÉ:V/a$M,<tUalE2_O\55=f/-+:dװ!sկNɃ+v3C-j`g땲C͎床ٮun΃U	8e'0-uK?,h,kdCmqFh#
F@i=ڇd'lcEw~?z0QLlCpo-r+C5J\dY
wnynm's(WL֏ֺ[s_l9~X{2Y׮ԩ$t߃3G?BLY\УQzhb?n|2s`}Q:iomlW\~m4w2暈4ݺWg?.OmyQۯg1i^po'o~YWX=[dȞ)%{S{FPL#FѯrTVf%1&%IkJa^"%+qWK퐽nv򟖅YTG*אXpD$XzD}D$I*&d_p-1ϧ
gLp413ٖ0ע1׹ע?]v&t0ެV.Rn2r\Պ2\(7T7XgK0ۧ%h*s}1!b\$!mX5|ì|
nY)f2ȖS\
b2zɻ?܎'|3~00҃[bɛVJs[b4^z9窕ugiJEc$TML*QS&CMy6qi1Li522Hzf@؟ַ	oRyzWZ5t8U	P-ZJhkKQ>V
wr\'7]Bzzsm|>+ReRlAlZn$Xfsw**(u1i|/^^ >ltIptƎiȸ4O
[󾜟5K
.2f}n^s3bl?u^4}[pbVnO{;`#tVȧAs/,nL(c띓Fֹ?(C[mYgz⍂#ignIWJqFWț"A©fT(JAUP7'QC._,im&ttKpȲu
0fY6}ϲxkǪ{\R9i.}%ۅPs'7xzEhjuyeKw6!b=Gw16uq`ayƗ]{k
w90#D!#ZcpHq(p
t45>V)!#b1Rz߸(5d(4YɨY=*@z\O.P:	RaNÜJ5@iԚU8ee#y	c?S~%[_櫙8S>&@WPõ.R
w/TM)%>*iY^
I,m<>&;na7nUIZoO]%KpT[`W6h>$~Ei@Adtt벘?\ST+ݧhy]ާń'N,c˅ڎYh8Њ9R0DEZ3<D)(AD"! Ӄ)4"^A,%`!e,edˏR`א٧`)JGhaoRU`N
գF!8x(wQO liƫVӰꗋ{lygN_ov'[b4+<22kPݕiklC@8B[\mKIM{6M Q=mi&Zc$IØ2-S&
i
%͌PːY!AB/jd+:VA4U)KTݛZ"YUKiPEղnUwui?sT-)ٕ]~2ge`eeSø}O,È(d}e2(aB֠2i?lD2y:9bT GU4rH1g!"BAgQ-q}>ľ|(J؃a?_
*%dñζ8X
E(RZT@1TE4z/EIu3VIN0pdE!AȪ(SY/!;dU
2ihijSȪ,9z4"uhЁ:ey^
f*0ժXƥd>^6]w0w7]NzHg|h¯2HOϯV-GB3V'^/0^
=RAm/(i״kWÄջXήUIZn6N/pe3ee=qnTvX7lX+4w˕A}FvB[d-Ѧޭy&eSqy7Lretrƻ1Rr2hSօh	@_VF#JÌڐ^x1!1)"h,$:"&	u
+%KʠI'mNkgЏfi^y&eSJ>X-WML'mD::wfwBfV={7{retrƻ1_Oݲmݺ'n6u(x-WML'mNR"?wfwBؔ)u}93eШ4O=K.hn,䉛hMCKNXn
!ً[gLW=2eqdځkkcWz310vsZHS1jפMnN
	Tn[R#٫|	<-Ζ>VO}{JAudzn
b%fa4\ӷ>,ߍ7ų;~>~֥|Uaw8|Uq#:eeʇq϶Hٵ4}/:Kf
BlA`?<$т{Ttϝ	ZAkA|RѝA`sRjbnh]:ݝ5٭U6Ūfʷ_ӛO?z&
|ocfhe>мk|5NKYԉn\IQ'cj0 b:bx{#qFDx#d}| STEdiک2K5x7h07tT&◍g"D)PV-lYDH1
rTiʷi%̗nPP1bTmݛUHՅ.Z7T=un5ДX!&F&A^c+QN+dPY)x3h.PF)=76lE׉
+e'GeSOP3	!Nc}]kg
Ƌx`7'0Pj;vdd6LQ:sHՔSZ޽{OiCX|me~`V
)Fp@3l00Iz[W>7iĿY޾1rŐ!ᆖi:4(
晠R	7-yWb+vb%`!+žg[n/!U#uT̊T)@8(Mf0lM)&$O@Ym^Emc7os;>^;6M֘0cW
x~
Q>
ΆcySv8)g<8FWwU]C?uwCb+8ÜK̏טwCfk[ceOcnYJsHSeK8#lzeQ.	FC#(kj9v.f+6։!=;+5ɴ޵oLv<,[$ZB[-9U"jhц#.S
5BBr|u&"kIXǒ!kMG\	ݛ*mƳn
^RR	u$73aF<8P-b|.VQ({hW`;Eq-t֋|%bcO/baKz:p/\{CqWu9͊I%(&j8cs2Nn;z;YT'\K)$3;Jf\-a7&JN`b]B=1-0-AVoq]'SO0ZT'SH+Vo;ͦ
'dK-%J}؎–**t:5WC${!$K.)A?͉NM0vzQòH44e"7C	+60}Kc@R
xQZ	N`~Z^)u1zoO/0h[}!#{\]AӮs-Ut-w_RCg֋cm
敡ɕ{ex&ѫXԋB2sh  Ji;	e~)2]wu1N XQd,Q~P_,FD:
BH;d5Ֆ4
VFkAېFg!ΌfmH-~!*)҈
fĂയwI$`ߏ3"n=}zAuΟL9aJ
SYPh䊮,$73llzE!kp<p=	HЊz-8Lb^^KO̻[NS׻{qM"usuFxݳ,Xn3.X)fV1s׫ߏz7^ӣۺ>ǵ#x4^ DNG"'Q^"EN9GRTKwlD}F&x€8h.HQ1UƸO⹍ˇqM.D{_&c[ףDAٍ/E2UٔŢHPK`
c5Y|%A%}kcILx:V!rr=Ydz,2Z!&(>YF`ҭG)m?ʋ'3L҂H?t	)=6% 1
 tT JsSYU-)wcqajW1
5kDJa\.5$JDҧA~GKynaur9閉?;ڏAq'0R,!IOf䧞0!kvO=V
n_~|X~<&6md>ဪ?BX?֡\BpL/BQ8w~Yz(4.,J#aunf$Ĉ|>|V2Sݕu#9v$ΛLl_8.0hTXI6$QEnH--g^
(젹yzR"?JtQ2iސz-rz!B(LxY"P)Q9r$+3Uݑب~یd5SSK *cv>9*dJ#))SH%	A\di:f%dP]ADR8=<&r[y=`ae3dg^&?PzsChxeFZBBNsv3@kc]yEtXɞ[<~Cu`=!RZv&ŭ*)VF8c̛gڒgzt͙|;ǣNuny:#f.Y!mZhylF7ü]C۴6.y䙇i>3"i!5
*A:@u:eqnfHVޣ$;'(HP)[SSYxٲ
)Abv;Y&ٱ)fՂECrPpO6v	k~ӽrĞRLzǓVY9t6,wWR(;bps`1Cc+38O8dnثL61nS[2dSRḑio_*^З`C{(OI[dFsl^}ЭL,rpsEx4`f~p>Np73V-T"*]8sU3k$Q
4}?
~:/-,d)wgv,C7{=pƉe(
CvLBk%;ɕK$qH.\oSs3gw9R%pݘ2Arp[;PV"'D`*|k:l	# QfW:g#i:M/rn=AF~ʱ9J^tkf^BN>~7
?qȹHL7G(y|~4@t:iy.>^-5m_J:xR\;Mwt؄p߀_ |(`$TR/CC1C)Bs(C=~2a+i,fUw۫7ʡͮ.v{˧Yt?(8)W\,(&P}ӿ_+h|-?gDK??$O2nSse4۲jڪgLO_$؟]Q}Lk)6J{VOo>潋uRjExbT$Ϫ=!؊8#N7ȖI=pϠV2{"Kپ@EŶĸVx92oqe9Ecըу"J/>UQBEOkULZM%QX>T~W~o	i_-| 6d &/pm@t0	L`B:g	?,<(`èX!."cȎ]vztУG=:a
=X]U󯤚@W{LżwL}KFo=>O1c~WN6IW)tԑ}t[/:D*"}RIxNO}5Egh&nBV
`-i!klɵOpr`^I|܏JT%8ӧx)6KjM_{leںf[5 [m6[!DzQV/Н@g?i
U^=6&`SEVu
E=|!cuq{sbB|.?jԏ֛Ccol2)& ja.
CiC(%`Bp`` )#,a|gu;ohH H>DqE}1e!Cnl.\_V\QUYAU6G~uu	әμw3yϼH(a+cЏY&t0	L`B	/FԵ胆"_Zh%Q-u{pT
j6gS"Y:TT]JRo"Sb#Sńf:/C,H+ e?BRL!@O-5o?ԤTP	duvE
0`MXԦVY#m9ڢ!+N|oo0)ȾMŝo~'c~WV~΃dWD".#xVt۹zHG
lPЈ`00
	{kR7VkT|ԟ^ać@`!a_jbK)CϘ0D"eD{)~/iއ6UoJ/@WkPRECpg5('H
Vej?U%[.d֞
Fd!uv/SO
vE[Eؒ|;NNk$֭N(+ڦHff/L}ăCes)j+;#,kW,h{ˇmȄ>R)[d3l&][F+^i~`Ϝ8NlvaxiZk$EgfSMBIԈGnvcK*\LyEhyw>J;	%IJ3U{CgzRl'uS]qKPBJJaJVfUlZ
,)>?VtN%+&JrGGsgPMkXXpXlz0o'@s1h!䤧v^%JS+F?dV)umhs=p$DڗَO,'n	B,Uˇi	ƒZ6hYf9"B;9m۷;R%vm^^{H;{a)Xp큆=q0w0*OZp]g3k3ֳYY	K9Bկ3 95$:JIYvyS`Lje8%ւӴ78$y|5U7>sk}Zq|3	,Ra~<aaBÍ(o8UJ0DJMD)njs1qo+*/`aE="6ȅ˪\VPmE嶢rʼ?Ud
UdZ7uZ7uZ7uZ7n<]Vi]h]h]h]HS8e|u<ڋT|j{>tڞO? F]/͸Q*aYG-
.y9>uU-8?TD:M<F	B/=0ZD~|DX1%ֵ
`mg){T@@jDk~mik;WK/<75yo{k[ޚK>تh!/ARr1y4к	к	к	P*vx4/e尔O9vgvxjsROUtxʟB/7Jmvn;V*?ٜQY<ӧI\jYN2I#]pq	><)2ݔe'RVD]WJ,*<f,@[SΝT|2shˆh`6u g"["'0Rl
ll|5S~jR\J՗O/Hgdv3fpkP0*J(uՙ%3|}G	&%h}d?_]-2'{ٵ٠v#Cɾ	8Gѐ᠄Љ("x?](,Y\WuBش@wٗE	m6/j#mz.f1{ɤK"H9](xmms9L[N+rj6r0̾o_v>a,CIWաdy+ԇqKq™Wncn񞡧s-i1U7A"\vgo?Ыig<<$s
3)Y G,,KF0{濌{Aɂ[ѯ
"3~vi4}a*hl.īcz$|m<\'*S^8Gq/L+k)Tj-f2%GMo0[hmqV0A!	yfɉ,?yΘ!9j.u70xBSg$s4_϶T
[BHQnu!k岴l0jZNペyڔ֨g\6	8Z
xlDLS	,(uģ'.-ChvW:M0ĭCZyP1.:e[}` iL!DT堪{67rjzylSpS&q8J+vXʀ:xWqk'B6	67]!E	)֡BKl($P*q0
sSYJz>򊩭8e:jP
^ڙF,Hx,vft5I2eK5?k[ݪ	ƋaFajAZZ}7Z/kplfGg/4d)$N
6?V|.L]0on/k>g3OASOLj`WY2tII=f{y|:|N()suҖZ0y?!'2͇RKHdzEg$>T-MDcld|0<J%bb8VgXTdG\@/x=4ux:;qdbu#)Wa-mrRLaa`=$ypy~
YC-HRTC-3jFxҹ{{I7]{G0w)&~+Efc?<h{\q
$t1Y'Fy*€4e'dI/21R=߶Nݏ~izTl$}K$Q"-{ysb#$ĪqE{	cPs~
VV^vϩX88RRr@)Ƒ53fsvN7&q\[D2=a$>ڍPWjpCU$Q5T4`ǂ`Q "kKuT+,2i0hQ!"
A*PAEBr243eْ1Z}
j}ӵn6M׵)S\
od#N/)LT4fMD`.|WxoQt?zvo={c.pUX/}xn5΍HGև\?pGnWtCY|q޹0A
(Huj=FϽsѳ|A0{`5#TUaH_2{a#00+1r!uI'-lN]dV-kʱ2k,gyldk҃szck tJVTcpbչresUzG2O	UEH G0a(R!p_Ճ!=*=W\)F*ײk\	S%HOud%a	iaG.
	`]RU1?,-Ȍw1W/gPX$!:e7kig{2kUK鳞"q:
@>=3cV=<
"K*a槸wD+aB)EHǹ##"dA$@3IJ8hH_15H'`%_xv1co0ߑ?)޷2R*i󪓬`er	qҦX:Ya_QPΧD@[ŭӼoo-}!
>~{\UWdoE)
lY`OTlS]sW
J°0!iaO̚-D>PJVEJ,TN ~<}JB9@eJtC'BN΍q>0zP.[ioWWدx.wZ,3U{GZctA\e(D  'X/Gc5WX,x1Smh9SɤK:V2C5:fb#ũbJy=w1O%X	9cuuEcOz&'P@ fu2I̬bᄐBV^/~Ur/Gi5g0qћaR"5naǟ5^{!.j4	8'iccq`V
r>2l2 X0	/f:^Y'eji)eh	h+(Q"(6 3rRNiEWЕ>tO0#{q8izUx©ө/%R:Ph]Tu7=ufM/:bE{R
wg~ixRnT;y(g]'>O/Pt?ͅ1.?l1n+wN.F}^Z&}*G"ߗ1ÍN>-=w{qPUwxx.I~.Ʌ!uV+&N܏V
YrݠX5e^ZQ\5i)s _ca;)>U_`!&.Tԯx MjsXAT7OԎfJzgDKJ$.#;nϰ(RԹTᅺ:uhAρrsqk6\KbDԤk7;YnA]WZ:SOZsphGiE]@"~)G/C]/VF15yzSdqg	<'}'?z.E4} Eij(zpfyNtEvTgAK$
ESݯ6B(!l+ʕa5&QmFcnX1jR3k=7io
586Oy!m>{M].C23x#Ig<|lN\>dWdF{͛^
r++\W˱w/FǨl#Kyá'#D4?b4"<DA$0(I%
|*~	 2(8y"Zf]kMb?rl|po4qnF|_M]{|[ӝpoz}EwLtv7_mw	!뿹wӻܬ5mRR1hMQ1fM4L)-('q-]mNNz2U~
lܾ5Oa,D.O㏸EnPekxƸW"T9br/u57Y|g-/\`?V7o
YRLXE QoQqgML%	Rns,hi}X'^wSy'juaοF*Gkx#.Yz+9k")RLB
))%)xdB2GLTP♶_&??2BdǗ.C)k<)ZΣ0U\Xb#:
Aʫ`4	PIܲ"`*+֖GN,+wX1`INHGO^%hù}U%H
Ҙf(0MLU4
5ʁqu>T3c7ʥ,X3QhxHKʊHU@M0ڣHEDItoߣ|qq;I2״drW]\ӅBjEGߩNfUa,*V&VQi$~h',sHD(-ɛ5jHI;bԚMC;O%1zĮ*gS_ZEp*%љ;UZ4B)TlyxVj/Jfm
/	=l[9/{Z9^3tmuIy
zY5è6w%]_~b|R.eRS	ml|ZGBBhm[1c)BZi<>;g!Uhd
"џhvn^儳x˟'tw~qu~w77{T>k/F#(#e)F?`%gnV-gH~ݎcqnxϥG
 bu4c9p-|N|)y!JV@	tk+{m?b0W
܀y'-~VC_NdoՒyA'wS$cg(]2n
?>ly}Chm?T/412XlC-ku:EC1s	%&˫y@|m$M'jN/&dr~?˅K
^b<>MCɟIGa0U
`?-D-Q%$Ppl6aK9} Z1#2:bLԩ`HE0C~QڈWѮڑ[gz.j_۬>SbَLqY5!Ekwyy}:yL.U1St>:.r¨>!'JT8y{~s:	S{9Ak&tߤI	ua3'llr|?쾲b'!=C"sO/6U-*O/w7!N,{bظbgv0Owf-'+V(ƍkV
TJK:jȣ)lƋ)tA@b)'SLZ	M)Ɨ8;9NvJj:C%#t$x0nJ*KaH?2wIs3bbT»]bԙ]Qb'O"F@@$F=" F9fs׶b|s#
gGۅCWx9.׊
	=|j4b#d^Q@gP&Pѻ\(8l&#U]f"
ӎ2	+TM2+Υ<8ɤ3V(Ug](Hz7!Vm=ID2m7]c}Fe,[JK)e6*G"ADtL&LR4;5ځj-dAi#	cH@pJˉv &zi&9(|RLy\EbH6Sht+1SO?VJIZU}V^Pmj{vQvcė3qyrMk|NvꂂY3,(1߽}Dy~Z
'$u
I8^;HF^Hje}@ҦdOHDd%^T]$Lc_Hjk
IψWH$?[~		&mMƞV(=``ا;,;,=EjlF%Z{$f[.T֠,}8^QjM-͞bC|]utK~K='rc~EE_7cms:'%~،όwf
jJeCf9'b.+Vx]lFJ
HwÏ޸|ﷷ֪C^2Y'Fl^ۦgNV.05~}yxPu{yY-;6)+RV/>nv-sR_WhqqBmZo7~XI1Nv+<heq
OB&B@>֙yGdu}~ZmH7>ngÆt==&wJڲYGe#3u0fn`;S/@x}C%@P}CлnCsߐ9fup*d{eא1
3=ǭVgpY[0"J(Uz(c@g@P)c3D DdՔIƯT!JW[xgįlXxyCdx1xdbRfY^V㕫5Aԫj+GGq3Wj11Td=|( oZ
ݺfK 'G=ũ}6rcG/3kS6U5}	3h5	
!iN+Q2ǣYY4ϙY>`޵/av6Wq>Ҍ+J\ŕcXURDB S0L޲5[O"+Re AH[AX	\EmϺPmPi_=aM:W&nԄpJ#.[\
vyXRrnh}ЋSφu`R'JXS.*0r27f'0B?Z?"1El0cIV?5!5FRAg
|JIFJ#aR2Qx)L[ȡThE#{%ȭes驱\Q1TEBc9Aag,fÚT3mGuLTkjA8w_0h`H@5't;CbaD*duS`ܙFS3A,(!TkښIa9)'=isF/;UÔLmmfɗ̪	%jyDԡ8Dw_~܊A6d6wi)wiJEٞr{18cs$,ѐnQɭFLr6"8zN(y
u"z@Tx.xnx30`znuFXi.{!ۧ^L1u>p4x>.	rs:
f4mo˗z3B&=ex`/^
Gga:,dCafcHEҗ'{D;"b =iZ(*Mz1iKH?|YT!#z3 \f
0'W@yߓ:TTd?b&g=[>,I[4ϖni'ybkֵ)=*
bV5}I=/ 
9K$+JiR,z:%E0CHY'4o=vɃSBA	[⹖9(dr,K %#ShV]ѳ~Xb*b\vK#]zP\rdf}\ N,bfc96ݚhߑwo~/]bT
 ^߂CD@P'/o|8Ζa> 3Hn(:/pz~xI>X[d
)p{?R*5~kL'yOƊ5CZM̹ildTxخ^@p$nXZz0UzO?oUQ
U?[e˼	V8jG`ʻ\|@"	)s%Zf1!C3mo"S	nv
$iw
Y#RVpGRA%b*8θ`)H_I+yUbzVY2[zM@^F~-|w*ʦ嵹CP0	nhs |Uխ
/؟
tJpR('3?\G(6Kx±Xr
7>iQ}@6K'%
Bԓ}bXʖkwBAMrXY8:KZ'uO^5cM1fd7ScABH<^嬂3+h_9r<	NIX3)x:@B?z-@nݜwX'#HhģSv΋C>gO{d|xv9ۏc37SͱRvGW6H[d<[<[;151Qv$3Y"2AǃَguLr/TR+">i`œVO!1d)l i2G0>WAHG\y5X	Q~N58A}@
%tx@Qq.P3 _q*7łwb\pk4[&84@Aȵ|JAUq&#TL2#	L>R*#ԛq3LxQEIy@iQ#$!Hʓ.i$f~L!D 2GNbЎg<}fL$N}YIT|58Ñ"ǯDSX1.vjjpDk)Dh"
t@0ѻZhG
`d=& pyg1!BjOpv<zaCǨiAc<xޔe:QH]TadΔ`aᰊBHa#`$ qR]HYjG6x\&<øm+,Yf2{TT>d$-8.#]':G#h ׎_d2Zh.kp~VϼY`
Ƀ3A#bU-b9yd&(u}hާ]) Oc♽UT¡oY"Zeٚd(Xt8y1J [
>=ZKYƒJB5=Gv>e4CNBu7wn=}t.tX/~\$e8~qim&h3fʴ!,;NOm@}$xt$nq15)s8ϴw9,]#Yaδ";n$n.~Z4/;7;3{WSv~tsnk/Nƭ.9AR(>Y`]'3gu~/|e.P04~pcy
kݔnD@M^@O	 NWRJXI1!v
Ai܄ҙ&zqxs%"T`AAV	Rsfr@#ВJ$s.2j׎oS1p.HI{Lq$Ŧ[%Qb*[Ś;nnE)`cvTdm$O<ͷ%ƞ|Hʨ3Pq1vwQI'<ڴY$I7=LbNE'.~v{X..>@pOosjt|>&$;̓W{U7KS#EiTl~3`7S%zR2)S ɚSTo>ժ㮓O4AvTbBO('ZںI2ȼĠX9KnyJ"E}K,/-@v<㙰su5$U,O-
~D*"S+\PjFm.JBBi<JL;N_}C-P)~+D[G_."2-ʼn(ox~On#R{mDͭeE싙lXH:C||sDӑ+9JpU
x5Z̔;;re#ю\ڔcu.y
#+:Z >'Β
&)l\ܬ<ლMihY)&Inmuا!/TR\`	&+l΍
)=F<Hh#2³:j;BY'cJt&clBF
-էfV@ާ}"Slڣ4N!j$gJ{s8hdNSx\@>.1Ҋ3
U4%~dAl)JQyʽ̛`5#8'#H"wVx<wQSr#TsebFam(Q]AP 8D")֬	8 DkL&b$e`1J
8Q!
I;5s{6ODcD2:7%1'^Qcj6#k2nлq/c?*E뽍1%=?_pKa?n2	Ry}VOb7Puo/<|pVv#ޯ&Lo0/VAŏTp	yD72OKz`51fKӛ!C3%?on`Ur`E%Aˁ Q#,WߧlՒRC@&h\F*"j@I

N$
6URRGdD0ƶy%ࣤ#uQ8^}јՖeVDyr0f2B<6*V [d;f!KX	G;E2ƥGe(wy`Zx#Z`i΂Xadb]Y҄l>@]2XjQi
+$pMR3𗂪d+(TQ+nZA*5'V0jİ˩ߘWh[&-!Dش*QxvHzѱƖkwPTg7t"
@k?wʩV
ئW;O5tr"ނ1\	sa#nB((*05Ĝ_K.
;1$V\E;4oo2WDi;$6X2At=wWZJ!
p	 %  ՛h8#kz|;CW{_J^ٳ?
f4O		bGGwL+B~ꃂ!%}>'PnM`^iE!U%D%({0x$ZE 2.ni(Dgj@P@Y!u0]Yo9+=m]ÀzmO=;Ƹ_];r$w}Y:b^%'R$/JP$GVD3"8P"j[FFKio%1EF@%DsGCg'ӥ/OjC>	>^Q;?Ϧcxj45ƱGqhF?Og<ӧhv[^_藣/>-0WJju+1ގ[1˧ouKq>c<l^'LY3L=VNJ,iEc.`4Ebt;ٿrhBeE.hAm_NૂP!e 
<ç9+7OlԪ0$z^K1%X=#${n3Ad~-+,JrafV+E}b/p))N!j&[#P̡r,H=>oܧ'9b!=XQGU֔ނ$&,vA9Du$L֊ҍ>~h@XpL0℟
x2YM_*ax0:Ќ"c$rVFŔ
E+6#R>V{ef狈5rcfWae^.g-.>W(u8ގ@i@'HM0vëB
tc&
t	Xr@i_9Hu1k	a-_
c@j#g_NJO;~
-
?]Ru&~$~)wў2X1;_{HuqY;7{@5…dZUTj8yJRYYcUkcm-iX\r\|c}2X(Yw$p)э;Ɲdi?Ռ{
$Q|xAf7b?qL?R75D?Fk^tw޺-@񷻥^;%mI|l]~N7_ѻ;0aV3GG8q;hFoN54d
ogc]#Z'fu[z-}33IٮYفNMqX6rmqs88<|>qMJyٽ6|~Knz7O4ӓ{
oZ.}XtNbrqi>gT/:(nbezE@ax'6.	mkRO]9ayMow4D-CDd@Z$d8ZH(X$	F!IqF9L(<]~Q@
WO*"Y=SgPz lkQ0
qijgGCP:?=S1	]?
׽=5h7wc`8zBCЪI7aջ!>˻w?}G;JٓxUg^ń(8
1WDZn:q'TeKM"q7oRʆ/?n
9ѷU>_wzBU&~YZ
!q˻fχwb \?MwˑY,&tvEtHI!"'IAL!RQ%#P|!beW(!:[g҇*L;*/\IZT	̌f͊f-뒐bFh"l
t~6ZdrgRAwp3j}SZ4x@>+xp„AvgcHuŹ8}iӉu-}5-wgfm΂@PjEAZm'?VioTErZC:wln:R]-.-8xM'g_y4B>B03ao~>	__?icb3>fFecF)+CKi7]
y<7ϛТF$jdHHٛ,y=^*g)VՠyR4[)uΑZ\A)Evdrldh戣9vjA&00@Hͫ N'C&Qc/<9~W4Y?=udFa?İTXp:e`/ܰAx,؇errj'A5h
iE9O1i!yaAaސN#Au{?7ofcp	rzP'v
P:45`)Klԗ 	5eg%xP>%稩!1襪Yȏ@.!kSٰ:h*KjL0s%SrHƛS^Cjlh%p5* Bf
D0Nx>F~bƘLE2}kXa
~9
^_\y^AwWO|w[|X7%Pltc+k8^ZdQ\“SHbϸps*.ُF?~,p5\b]2]/Պy%ˈ3i,yC@ZH3k%X0*eeL/F55IN>|t:?fO {?]>ښ]iX/~zsR_C׏_1K'cg
b]܅%ذ|)|
34sYJX+UF$^-7IRr1*TDž#^zvB/LZϺ~>IKq]&j#y`ePH%}i*IkY`*n$O"JSJtWPA\2/VȆ)a$1$59I#,¢Xm @i3bEUgXӸk\Pc2tN
L! `5ZZ\m]p;fw%?_>Jïp
dasX֤8>I?h1[}an'yGS	WXW\ouWXH!iDQrzą̙z`EXf5Ryblcs}&x4G{օ@C^'0J^'
Z':y>׊3NCp=0@fdie0@U2N=UP
4$`Ez̈́PeqD,@p
Q0	0@gfLy	PرDk,:Q{(&R9A2c
vHgqzZ
b_kJ;1($k9(7 J@Ky`\c½$mvJ)2z
D'J(LuASSg5iPr< #ry,X[~kQ%{.R\dm<뷛x;M	tfyh~7]N?IGLnb[chg˿DpvvwF2F\!{2_R_&Oެ?H}Ǫq@/Ų;N$^bResF)͕:Rƅ
&}Xk'.*OQ2D!a
DBSb_PxeL	NB~
c%fLԋf8Gy	hXiƢH+
Dl!
 "Eo"RKxZ1蒛`ae)u;"X*=&4RTB@3<kTo()w$2gk6}!kn@#N
&sRkQ.1p<'`Y;/gT>Xh^pX=X1ؠ5A9^\6"XU2/fR&4~g4>3}77"[s|ZjcƺB0x
ܨ_4r?[99x?n>O-gX8+S^ͻߋSQ2Y[L}4C[$w?J#ֹsFbcD]qrӛb|.#Ոkzm~ɚ%2q렯	bǼG2oWfL5¼,RץTй_͐H	ӲԈ_:,}4>&x[}!\&ig12:L{+jjy]JmN}{ļ/
^mѾ͎ט",VO:Tkqzwn+o
krVT~ϡRzj)25QyFgH!7TA	RDtfqM/=pE55Ҏ˂__z؂}x~)|
U1ER	\'$R.-$hlL0cZb :2oTW#?,-Ҫ>z/
׷9ø
#z+ʪb~躚D'Md7jMVt6_7܀|oiͯYx	D#>xm4Ա::ϑ_(.iH';~}7I0טMLb򸽧k؊j96Bl
1JõW>8
 :TSW"4+xԇhF}dr&y5Zí#Vka(8E_ipc	逓	p
xX'58!tzXm^Sk?SZdE$b%i<EGƀ1Ir 侓';-+.85K%8k
L0Ubگe5ig}Eeb2,&{7:z34+ZK0WiK*vp ~^ô$-7ZZӐֻE?a4Z2r?{3\\FLmVpe#VQwPvv{]קh>Qd@Yt}v'j[^Pn\kbTs+j2%iC8?3
+zLc㺫lvWZZzP[]	(UYy*J"iWS g+Crgn3,loO0-9F<[\r9Okzp]Lf0எb%VBB4ro		l4JmZIsNmF--{u|_on j?'Ct.E ,0\;KF+yA\#q*(,Nm4U<*=%Df1
b>!{1b$3y[qNp!REkI':U`!'CsbPczC"ךCX&5ASdb.'1e9[^Oy'7YT]

;@|y8u>YYǶ]_oj%kpri
Wy,|-f6\Bʪ==镠O?W@9L(&j7:`BV;';-k*?^}t
VH޴Rk?ל!B|(i]nggAB|ǭ1;tUI4ǻULK׶$	mnf'u?)8$ʜGǘ^[0ZEEVRN
B͜6`3VHCLXHVR_
nczTdK
J[wvlRlҗTc&YZhfKY +lyMk"MR/&]>˧_~JA({f	HA::ڜ~4 7)yk}
-!!:c[:mK0D=>ȁi/Ry<<|6mKl#G^?Ţ{8nF	NW0ٰ/	Í:'B{"ti4gjoϿJ_F]˶Q6]wV.7q9AZjub%<~пeݧA6'^:vNQVtD9M?뷹]L)v"y$zY/oTGWxkoc(|<z|ndja#+T9HUh"	h,4(Iޓ-q$+^t3]G~yWž/5|}
G	k{1M+?O
tfhw^ժWح?m߉Z82<1Gãծwx ʼnWL	oZ'`ggP4N2UE!G(Q]^ݝvS!"g9;AQ&xڕߦAvK38¼,mI?sk?Bc1qjU=_@GzɎT3Hƴ:T}j9e:U=	Ւў+Ќ&ExBQ3p{kZu͇QN|ֆh,Ci
eT@TRltY.TtPaF\Xjw8WV{ZKx/d4Z%8ÁBk3UfUN޸o[lc#CiM#D+9-B)Clyr^ji㚔V1]zp.
{1!_~ve2-üݷ;Wfx?@OITntυza5TDZ-\)4?qtw(j(#⸪~[dT}uf`:^,]JK()PMȞ~A47[q|ԣ7%fұڀ-91ىGa2w?G?8mb=IMf(^M;ѠWh.EQua#)rbA;Xz{9VP>RytB+@>GYs)O@c))=M̿̄dz,*UbӫĦWu65FGA6R@VÜMHkZ	"!*7vV=yn~0G;lk~+}KD+hf.ǚJ`+ě7QʔL*ez>1rgή?upv.TKZb\7yΰ!EY	艑.G?-?,*M't6TUB8.G	5|8]g>bf7wvq${4M.nDw*!tsA
eQseFW-ъ:2!hMvw.P!NN_/Ҧ	5aQp2
-<f

̗G pTJO 
"B1FO191ڰ^
=L9ۃ͔U@EFDMhsePj]>朤J.sRBbm9B:7ܜQJ:TL"Rv}M1ȖY77a?I&}>ofjиZ&|yϠwppueA"Z
0u1'z)kjAtRDƼ^W]3!?6E{|!puY$ږwdAWNW	IIozDGzTIS0Su
~׈{d^P Dvԡ'gȜؘ2-)	.&7~ZH+>2w?pқ ؑwcp45ya6z{0w-@WqA+k@"VihE!JRSP,PntDYc]جgdl5aJC{NEw{B'	P_WGN;S:mly@%vu[B..;6vKϦ߭;}aX#K) 8IM(N+s u{Rx)cjWh-EU	KVZ#nSho&lZi7.zERT)R
U,X]wv)kZr7`+Hjh,AK*]bqBŎ=	]3ȢVGI-Ewue3<ʬ.}9ysZLS`1911v7++39|=}_|}w;nlqIpaP^r?)q
'!v~;?A]
JJ?	JZab
\+Ԩ(`ǘ/p1)Gcu7zִ
9T?XjOLUL7Sb\7%o{h0&e&7)>'w7tJ^l"A)_rF)fDzr#;M?y{QW?څ\y{9;}w>XXʹ52Um_7q.{>" At>bg0j;JM꣑*XX_ѝm&c5㺢V}u.zZLLx=~ے<:A$kLuPfY}(	Ah5&
'"3#$q(:#2@J+.k]
;
MFE;8Ǚnh5No"tz_vǸJ1Rwʕ4Med΀U5:(>lj
\Vߴ
#m`:<Ƶeu	gjEKըیuNtn2Z&PD{
VAE,a\sa#F|QtvkݘƜZY3K[jVh6#[O$9A!
$&w>ߡY0D2_S7(Wd6GL"ƒd	kV&td
afrjE8%T)Jg0|i'(`QTw8GL#ǃ= LA6"$}A=8EMcJAQWhJ
MИ
vR?+]tA>zMZt	!h@dTC/0誰wk[
vE娂y8J˙:wy8@'ц{d}:.fү"@mk-%UWI\%IpUB%1K\SQ#8!
DH
N|r`iPKi+_@醳4˯Y~d_^54
.FgJ43mҼ,Z#u
BiݨvZqPZC	נ3zv7X1zZd=m9`iICd\8bzMQ&m8	6*gB}((šhz܃k\)GiAab,9%4d.#rQZH*J(TDV	sx_4B%y36Ј7puNy|wQ.@<:,x˵.㥏
%A	RImS6rxJwĦQ=@RK)d-~bY]/T(U.YLj|{QNb7w{lܶ`1~ޢ᧴ mO;v:"~'`hN:=_-k!׌#Y=Ndmp*"3XprKlMZ}7h,-kG+ 8UP,"Z^0RH[	S*GO:%[C9d;z?P,8$b~ӱ@^qv$ԝN5ӨdçSEu!5+7F('oa]r5*
3E ^Ʋz}͊N8MF>>м)j@Qkv9xܧkU~wPw}ZV{09aw"ō\tBhκJ!uTԣ-QZ;-ck͠faH`XL*y2L$42hq}
ǃ9jN:Ԣ n>T1)޵5m#뿢۞:3T!gOawSy)eFVi@Z*H#nv7y
ዘї:/ÙWrSm1o_|spkϸg+&]G詯1^QR$'FEW{<ڟ9ӿon̓e_P"(HhN5⠥zu3/b:ȱnLS[≦Z14+Wdh6cٲnR*Xu+ Iu
eN|֭tm
UtN	LvRi:Kw|>2Vh{~&ԟ_nX)U%)ݬM/gE0!'<~7fpL\dǴG5N7`hX	9CHWW}.#	8D[?71jss"⿬~G~ڕ?<>Oxe$iL'֤WrW.a<$;TF(.Q;xA^N6uOJo_-0^,}:Z>
zfϏ0͖J<{=B+'l,UtO;X#iWrCk'%wר-&~rP;8r_	Q;ܟ]ǝ}ӛ~{V]]}||g=O4Htce{*cMdf)EbDZO-OP:'Bh).EdRLJ9NO2F
HfT-!!NJѕpJD]UE8_1k8TUgfPldRyZ2اhh$D-K\@	I/L(	5f6L,K2R:<Ґb"i<[4	>6+3I+jn'lri4:CgH2BdYgL!2dȉ^$ޒ	IVԃXp.TJlmuH5>CPE`d C|m%L@qebŜ&\_E0jP:^ERVi3qcY
3W(yRxw_*'_*˳\\x7Z33÷/^Zm^~Xy˘K|yv(!ߟyHMaɬC%T395%1RFkЂrݗH"pFfl|膎	ZpÅԊ$S$(Ƀ]2!t*EN~%8ff$r$Lv5auOJhpMG8צOkN1LY$uP=TpQ*bbv2&V{_i"[M[D(f*'l-7|&mr|"/hhԑmIoa1dd3pABxkFd
axK#L?@AaVJm&|}qHqZt
@æ&97DTfp!>͊]}:faI
acG\K ^])Qc@kiz][YƵ'ႜAmE0;TepAC	d&	V`4<GQj(`TRƸL/T*+
ԋ{%\K"*PS`7>⦢VPC"0Y#dZ+B	VV"
m|`$Xe-4ƚreuZҵmvyE-
+i:j/~"k(.R״覄 ̎zͥpE1	EgvfpDޚVAqK\s:qp
LywHxĩ.jA
'/;	&[U8!rPt",D_/)R-JތTg0[zns#%ΉTz5>yvN)&T;;'T(lHs+ֶdF1):dKEJжme:!ϒ{l0
LFgT=
	[ݪ'wdz4:daBR2z4NcZ)CsTOE3:D&[n'0|@cK"='hJKyeyb6[{)}!jYgYAx)FmrȴOfϟh_WzBTSD/B:ap9TpV$F$3~
j-wBQ|q`hV3ocQ"	:ֶef&=@S8:&0AKnǷckEli=]dqKOnqXb6gy8HA,_KhIms3!O{3r1!.)4xa۾RĈZiÆ iQ0GWĒ̱k=cS	dj0Lhѳ^ѥDߍG㝴T+RDU\rI9m
w/W_s凈IU@I=:F4:䡵,F/K	s(K
v)zuRHԫ9U5֛xJ5gDĚF5FyĚ=:zRp%	K8;Oať=w#45>~DuLfq	AVW,\r	,VC)ذvHiKQWFN8ej*9Dp/PSqGb'If	cx֘;/R{<S >M=x(==s&OA*FP)%
ǩTU
FV^C7AyQ&}f|9cm+o*y'÷kH)v9gvD2!"#3|֑'dLHDDox
ϸ`Y4&9GL,ejMFveSɘŢ6Rԧk*,hc#O$ECLwop0rDyIHߨJT7ۨʐP%|U-KF٥=6%J{
7TUӒqz*s!ISAS9VЀe2@!HLY0͂0h6&DP@UJ5Q$Tael6P)8j]ZJ@!
m.h
/8e#H$1QSC(?ub5c8k&dôJ-B`Um*R
-्/*h,5p)
hRmQq7jvHh^_MF,1:rxdo.61h۳_l-uWןU˸Fh/9}؏#Ǻ0Br5u+ohА\EtJ#ζu'nN3r)WZ
>u+huCCr=)Sz-~uiSS?,$YS[{:LQFH[1x8SOD[	&c*|U|@SWPLlM)*{D$cWT?7k~%K* z*2[	z/r#(OWkMPųHKg5|gb0F˒,+Oe#VJ*WyZps0hRqրe2p@NN˦|W×	y]σ!en&?G^G-5&RGD_R9Цߛb@-pZ.dQXp))k>MDa멷-14+W,O"
^ZRRT9tW`6ںw*MnchWE:%#ζu.b:ȱn\E(r6VW[Sg쮎3Z"hkܟVDߕXzoϪ?΃<	D0J"Q4Scd"4NSTvt(
D(%MIF<T(3O-"U[2"4((SIvkp8E$.-ǟS'oJc}5c0<2/8o!BeH).Edr4gs(RԫGEo6ih0u[ɾ{|bҤ
 ӂA3MWqZ̘l͊1,a
Is>]/?U-+]'.^u·v./~cկ}D̓mWFX8N)6k5.%FeR\D[)A*>>,ܽpԔ?)t
o)y)
m&)^Alw䣵jGِ߾.{3Jh8{~ :L#KS.Fxheسf*"ryT6
}Sޓ잯^I+p0K8KPgMSұ:hkk	kRJA1I!_}@
/f)s~4%12ww}{ٴc"}zIB8Y/=ǝbXeq
F괕hjV)) CFN^#RٻHr#W
zFU )_vX16ϑcdž됲.̬*I5@TYL`\Fl:wmVkZOknl!Ѫv4<[w-{w΢/)ig;<'_̀[^,LyLDq){[g4?U׷[T׷⍵+7]*pY8kIK"NKE:	
H!g!e,Rrmk߽߰lw|*HK҈QXA@gVDEtD]9g6sH&'FaΘ2DZMCߵdcN[/]F	r@q
Il9x;B/OY0̈́m0<aZKDc8V:r2ΌRMBz¸yKJ-P:`?g)~'7VHRYkv9p*}ԂQPN򞠕VN9L( 0>nhdX#kQzRMq:XWcx09C4E
~uwl~e__~)MwǴSYDZ!1.%4?7CXfYs.7\'AJ:s.S&xĬ	R/˔,ߧO|۴8RNNGB2x7/wwV/9k@v/_\9)ջk-w`'H"7JF؉HLu"9jsQ\lJsr&AA%sQdb*i(+Rz_-9ȑ$E=xFT#4b*4DeI&H˘+"ؒA;BU')=d	!lL*%#-faR)#WD
'JBWt@%hIі'w2&dVRͺKoiJOH9Ӑ!R
6"a\.HU$}ԉMelliSs9I~\}uV.৳Ո?=3OB#M?#hm>$>}k"Vv:wzd-'1>n&NH;2?_\29B+?Xa&J..ݢRKije8)'BAk>]W,kp(p 
 lǜcaLE,GE^!`֐a!8-^Ĉ+#-q:#bD6hJQa0"hr mf[3McJ*$WQVzfuh}i!Gk3ў_@ۨC.VZqDzVvP>㉮&+ècacypLC	!-X[;xcB
Ș47;&rD!l50e%mF@uJjƎǨ3VYsshY<}	ch}iʲqe})ŬZo֚[a7qcdս;G=Ўy! Z؞NAaLԂ|b/]"3!CMK&3' Q%%/G>&d	DvKtK2mB[0[ 2e$R%dKAt&Ke''?)wDm	J;:bf_!f322y-$XOggkr-X}E3;sZTFb^%I:&g8@d&aSʳ`_),Rz2VDEJ9#Y^ז*90g
F+Fu6;(na"bEr)5'ו![3YQ8m|ᰌmK?{Ƒ_e@wȇCw5֛͗===HYjRPaϐ q,r4z0TZ%[	2L9aۮw)d2F=4=-t0n*l@O0 hҶH˗ۧwq2]AђyKX
Ԭ!mI9c!51CIV
X7=A$vd_jmuP/Bl
iύoT>6*SZSpll575u~;$WQU]]IbTOBr+QJYRdRFWnR_#[8~hGΠ8T?DUֺ|_#>OB|P:Xt>I~)jC'{joTLɩ׍p4j`
o_Գd25?Bː{Hm
*5Mݛ[g{ED"I_Cv6Ӣ	A#o7D2n<	y$mavM#e{)hzW6Tc»څ
g%h^	
RSʆ*]至0dɖݑb
gq	şf5fockяgxUM$7󫗲9Z4+M뗲n$y=
ҶFYȭ(t澆$'`%a
ĆZT#kyB@
y1ׅ<4ۉZy(fCyȫ~
yo0.FE"T}z%f̖rr;y<ݩydRG|7g::@OP	RX.#{!5\|2:}V0<{t7/?NcCH>wxX}=ﰿ'gN~\wrt@_|w_.?=Y@$`DӷҬ齓{OQ+RTxdbNwzWHDfV	cZ`ϛb{;S&r6xOk.||foMz􃞵"GB1rF6G_[OmƲیDrEémzB1r7'[CIDXr)K]ƥQ!tCeFDf\	gLթF}6Reu~Z=|othkG3O:0h!!1H}g21&}ԩ="$t:x;11<:OGQnN/L(dn9eճ|O[*#"8_l%m+ULioaSzl%d=en]5~m_!)|"BgWwnyҠ:'HRDpxIN L?b[Yz4mn>{9}O;$By/?TL{R.]++fjƇ+	j][nj	VβY
k1JuMlZf6')]8c{=0}'F[%aZɗ>rˎ:n'YZ!
 fd$C";b(6Rx~E]VRN)5&M655kpŕ8p1u*\,Kp9Ve
Lʗ$͛U$^ȔM="2'HkmjU;tpA[B
eئJK1Q9aL"3s%zkԩ4`iI}+||]{0#̞lse\#r	x-[ڲ>`dç?Xƒ+$IK4#@J'T[b׉2]s2]d0g"S;GU-hBr*(	$1XTK>& y~RߪT˜eFcEgOf!9
=GIbXuFA`Q<=vJp/Тi	 zW
ҫ}G#gr8eA[ts;dsp012J}Sqhhbi̭ϯ9;:ȦKֿcdPp_?%EF}<\<|Qp_9̩| B\Y!*s tvY\QݷnT~7V0/TD%B_sDŽCPH@ZVs&bmNx0vao[M-drE}w2[I.,KhfAuqy{wt-s9HJ9>(fHG@X9'Se0ScO.P4BjVy#H3I
EɄC
Hx~jiKГ\H/XH/<4gXŽHӚ0>pM*D^+pA9C; -]ij3@Uȹaih/}-`%xemxk%yKE@`A2Ƣ%-a,/>654ΰFc
DT oZE)f 3 (K]dXe{$s=+R{_l\>ߟ5ȿkrJ`	Og21#%{>Y\V:.y$vWjY2QW$oW	ŖLDܦP.K̡>"p#&0}6D;՝?Lϴyw%L
3twn@Ag['R0L
;P	;\
lt*&[#V[x=-z^b<%bi&@Z6HT!gV[T
{bhL	F.;f'pkiCp{ܚ뷵-#k߽8//ܖ=O5h!Uܜ91Fzԓ_=ѻR3a\G(s$V=Q?{I0؋6|2`zvFnTJTJf2I,JE~F0ޑe-B4P)TjPwQ>*#MM&:
E3J 5GC{S[0$g9gbtzC~czcUzUN>xNʕB|##ʗ(HꁝIĤXฉ7ߗ8Ox\,F{_	gT`43AS,&*)qf<`
9C[PLCc7Ao.8O)pb,,_4DG;1›Ps\H-MRt?q
RoK	]kf;>j@?U+>)ˣI5!"ۧDNM1Fn|:g2*r)wN.Uu媕Ut"4Uf6
%	J `=}kZ:XSx@f3hH=umw}kbtEH̒SRV8g桋5ZyaMG >Ps1iw9~PB^!N
yoñJ$e?~IpU_og,9פ8߇8}=}@/<>p24?Tx?{9}s*xΟу8"t"q!DV==nOrikP653r^BTT)̌+`ҍ:xʹ9vR4%L>04N)Jx+&*
l}3=<2.!ǯrpyLAusgYY<@8jV-0ۀnjuVtОfWn˂rZJ`KcQ'1?RChPHݭ`LW޺*/K$ZrK̕hD zvYpven
K-Gx)!$pӱ]WdzlX+M72(9;G~@SN{Q_WWe4fl3Z2P
2 q[ީ0X/^]y4;!B(C 8 M
CcKLz{=z{#j//G̸欓t`XEDy>wkGSH0æCi*3DꌤрooxZ-7vzD"kvD"9"عDTiC m
#SXcG	sv0@⦮TJfV+ۏyMJE(톒OBWk*zN{{e EjL7ae>p.@Po-qbZjT
ym_ΟWwcI>|	S^IWO<ν?Z2PZEWMqV;Nۦ#d֭ŀmg@a]ipظM3)"PB%LD0d
d%3K9}=s3PB%+듳hi<$CD/?i82Wzw/nξemͅ•Λ͐ͧ﷓`"[%(M^.ح/։m"#]\
8%)B)'hBnb?wp':=Zݮr6>LWZhlYqcì2.SNNBps["(8ٚUwݭǀv\Y%A"@l7l89Z%}\FwQh$GDq^[Sa?v9L[ׂ
dTrmJPyTgkŅ@
wtĥXŋGs/2j~lxu]ڜE4'pQ:9& iRqo#oxtx?9c6LHĕȗ*"59>bA]E#@7քRߝkc>wMLVu"<L4ȯSh*[ rrY198Kalex9;-S%iT9&e@W2.79^D.x$`]D|Dj4RWGw[مޣYNjDV;wZ$1W֖M'0^y*iB	XjD*
 LTᖣpG4ZWOpv~==;-)WϐQT[bGIm24đau#R#y`_JY/
pMҫgj{H_b~^]6YI&5؎mHc@fW
{fzyjS.[p]|U0.Tʺz*톫B$BUYi٤ފ@ǒ(tTR$40ֶG'39K\0U(e'ǝaTftǼbKTp<,F@N=˝n6Y莦t;P{y5
iDXXd(S,cH1667N{7Lcl_bf8s5S,	6pƃxEai(i_K7*Ļ3+_l3&,͌,:)y%c@HԈy!X<
`OJ)ƤwYaAH qZf*ęUm;K
Kn	8sW`#d%n+.Ћ7zAa_'-آzyDu/vҟ5MȍhaafOO	!Cc,S-f=oiXKDg8Ʊ
ܠ"ÄRD&qʼn#K)Ujͽ6:QFkI96Ц9lB(Pkt`()yk@>6PhS{thҍ/?gxV+1U¹d.]HJK+
aKZ.5s%Y]0=~iˣۜ&St{e8UmocܜiنJӲ/o~JpĉU,dQr}h_FMN׳JXUUwf%j&VݹEE̾|!kZn`nP]3Q9Xbl!mEdsbLL4p#D *rVGBH`=Mw1[Q,m6Tʜm|kmf4f+lVULYG<棆IMp
ò쒑"22A
aas.:Ύ
e/1;(㩝V:iJ'T餕*Nfpkf@}zǟ8Rjpb8ƀE)QHEf~(\rPdKzR̾:$&|Zřx_ciBHu7O._]Serxj;džpф
Gd ^Fk-걡`m@@,b=1JJ	8J;	nzw)͝|m?^l::{霆&Ez'M@
.9y󦵽rr5GimO߽`o/wZov~co5t׽v}a駃=ȽË귗OP#_yvp9yNzL*D:[kOmd)Oi1m&"4vtBқ6ͿF@vΛiO733x*>#
].
`[+6Om~^߼ϝ\M~܇~ptP_—]xV'?d,\$Ow~GG/¨da˥/\?zQs9H&]\͓$r9=?L5Fy/OzMg;tara#X{o6)?=;l}qmwx=)]ly=6A{;xAǺ=ϽWݭ_WG{݃^o,/~;
ρPҭN鲁32^x{N0\0mO@g]D[ֺw_7G0e'{y릫&:~nҧ1;Sÿ?l|2
WlJ`A[br XrHQp:Kl?}8xI(nCr)iqg{vF_n|%2>tvݘg,6Jp/A([5MJZ)}V+jgu"H1fp.0*X`8&1Gl=
|~f^Rܫl%Mkɇ^3)r3B}
>E@uiaN0̂kUDD5WMjUn]{{%4kLL`@Dxq;@P{%h
|Iə46
)u`(݇Nk,XcXpXp[4%K(Y
ER(JVEy@p_2?pja#E/9)!O.^r=wYSKQq'lpZ!PSFj))bSωKp,05g9c,yԜ,5g982ne,?{SF\/_3%B4A:Gr8Vp9Zi2zf%xcQeᬔZ@A'Fo!8o*PС/p
*b_Su~}sK甇)MaQDH:2DkC#nj)NcԑN|Øfrr9sVSKEq\Вa7W;[ϧ@X/ۡBl>ٳ_~|bkoXC=\~QHʡeGЭ~s^5Uɥ$F$ɷd~R*4;_g|~%!1RJySܞj)˗_0L3!RޚY"K>&5g*xcY( D@(N},VHKqgc,JHT*lНAr	l% g^ED#+0H
Et\V^S]yDŜ	S%fb)΍ROXzWO:O%*/v]}	@S}''zWb߮Zq؊T0쐘!ZI=k8,BR!,BR!Xie%K(Xp2f,EL\A{$Ia6)]<yPO"b%,<#NQ=sN'īx2/Qfkk&e5ZcdctV&a7Ya%dtQE{tIQ쒡c"cHmIon2;	SJvlPy{Uaޡmd'Vst;S=y5ھv	b:(=&^CEmQkn+涢kn+bm/40z%x/>-
.3?.'s,sE|2ĪYp29Grɐg8\w@6wg@|փ3 \6mT=Sq(ۦm\-]# N@L;i'qeA9_@`f>!/J@1E	+}pξmDl	:V3jB}ck/dj|u_Bú1fJX*;y6dn-Qp{FN#X3#
doмX<|i}wYz;
zb		y-jcU.,iwꛮLȗ&TN7-]ܜQ$miqkBmTҭ?(>8dܚ:n[gZd24{|u8b^|SܓU-'ik9Ûǿ^yl10]'F%
FQ)t9k'	TY!'݅3Ok0Cǟ˭Pk0CmVui>5ZzmjJ$W
77oRlZ>Ey72\wzӼ,4.94#Ic3==4SNWxf)|B㙧qƅ^5.ԸPB5.=.Qh!c䅿\-B]}ќ|Ji#+ѲD{A@IzTh#3>	3yǣ+5gV)`:ރ#euX"
ES>W>[,LH9䨱Rrx!)Yc2 Nz*c$4/I8y}1D&b4c{">Y~jk|
1	[^'ŅFzdO(a?}ie&ߵXSTeou=TD6
}[*R,)ɾn`@YqƁ^zyڡƁjq-rt3rSc]+Pd?(BԭDn([ǁZqP~uW/=.7߹d.{WPrs;NKili72!t肓][}T>hIJږΝ4IX?OQNjwM;Jx
aTXnVI 0tpq^v‰oi@:?;b'%,ԕ<k}
%PW͉f1eY1Hfܣ`I2ʧ
2m'턽O귶vNX	k;a	{N6Dl..BJ&IHA6вC(ަ[7:{٨7wYhwz$:gV[qFԵݎZqΐv~2ICEZ(*W]51:WHa)i$PT(JyʸCT9h'qL٦Ͼ$}~~Gvg>_~i]u;%t;
T'Wi;1ඊtrLk1	RfIQBJFF8ieg+Ӯ{7'mӛG7wڱ;vc7+*=;:Ͻ['O-͖׃aiԦi^|Ǹ=3&af4|ـs;};/\|\<5R)1$Gő%ua}:eKP[gyT
e=?ӗ)e1D&$QqUDv)ۓSfRL8,Sd23c"8!PLY5"(΅YWon0W{ȿp?SMO#EjNlEF @ֆHҡhL6(m=豨,!IZ:f,nٱm#/^KŻR
@TB$c
Z6rmE2ި!泳]tw4	/wO؄&\4	M,:Y,N&_>E`,ze`,3
]'9$:WZ{כ;CzwGAzni];󧋋˃7ù0k/+HyE	oJB4y?.HL(CJ5Y$eAc.DJI?FUw2eJ9%,1V8/lh}8uu?^3fzww#>pxouzaOH#8߽FC町>{vBZ){_z*P2A m|$g5qk/Þz^Gӫo=e?7c1Eprr;v]?ttFmN%889/bNZE0	>b֡J}bFK2v-w
sk˫to'Um`,#TqbLbfc7?F\?9_Dž
q@ɾuy>aC!TY:*!FƖĄQyK&0\DR(X|*gUb6x?~lh"dg+rO˧zy<6fx5ɚ,eH^+h"]v2F)vetv99:;$lHjM=`P$Oo>B1Iʘe\\2hLTfA 2LvXs8z$O础-Z=skU{^S2>lh/d sbij2VL+DeYD?̸D1r0$k(2	1^]W(O	V?99zטD	AUAH%"-BTMdb8eKdv4c)}g`u SbuY.
T![uɧy:52	cmHA$';,)j˼Eʀ06z`+WNl!H 8{y9fp20KAu<z	`*$)1DmIXNj.<'CS'Ӓ5B9Sz&fƕ",?$Y5"&諥]oƲWPpsq]M&	ri%U̐X~;lm@?2.
Ge+h:iT@~h|rxR]yY^VTaY٭`ㄱS7)-$`C|dGKd~cHG4E#hZxL4=24
n-	?Fkk@)DMn

pcLU7lO)L}'.Q2*ǥqx=XFaIpE`u`	XC4JKbE]XjwP>*˸Q-`)n`h%J1Рv(Pw`oD,
ŷP,&d31Xrր%BPNy p	Ѳ6,o+TƘ-`p%ZG,шee'2B^`BHo~)p0:
F$n`.
FAJӉDpg_e
L/o+[J)e5.qX'!x!nX2ܿCHm
[

=PŵK
n;\FDCjrh
,tՉ7uw
]rB
Kn;qpA;\
sy8CF;qTײ؈5/g
MmVBzpm
]Vwŝ`Er3]$9|i06]C&cEF5aQrtPNd@?,VB?Ew΁bdA''bvpK+c6@#`S)34g1gAL5͵ve2DdGCMlbDshr70oxTduR܃?w{#L}[ERa9vnF^b|)2mU%<7Jtd'"$1$(}.VL\FLN2ȑSAY&p;rSj\&&b3MB(^ob*\1x\ZZXSL`H0ViK)R o4Jb<S}~qyqxZƅ$r[INml=4〻l{ʌwehכO$쵦LE@p%GR'(PmK,bj席[MQYbgFRLa	:ok/S&K	43x\EP
bLJiT0dPReOl~1@LH$${*xdTUD0qB>XLJ8pZs	K3?)uN90/mt+y1BhXJIԮUH
fKd99\9n#Y5hD$s58e\qF
*mpH	`[	AcMbf=SmL1O"I",OqOJpt9+hMBpդ`:M25p	6#+SL-Z%Cl|KaJL~L')y"Q&p(ГVJ丁ɴYA"'o2bVH%FZ&)iaE껹\p4Bh"ajkrInSDM`	Te`)IBT37"CSErRI
U9X*pu \q2rҶA
%8;w4u2NLdRJKp <_>ᦴ"s6s2IKp@[G1.LL-b?kY|
*tUi$X`򋯿Ľxym:X9͇V1xKdBe/NNzz*9)?8NF~Q='/O~{|rx˓={ޜ篎wop';Wyg9_o<777nؾK4uͧj4/zhhհw}NAYǗQϞ_
dczds={68^o*,v<~,Qb)s4vKcU[Dw~.ta}5_UR|-/O'>b:#'!1*-^8y1PTTۓfWc}bY[$qQ<"r{~t'`zG;.6w<~鿀I].|x܋{~~pץ\5Kb9+Jׇ
.>;}}~ra06?[ܞ49cp*jج8k~*lN0tKr]|=̼Gj<9ȇE-JB2މ#7sh]a1w磏CVW3|Rp+t:-~OPj?npn\}BQֳt5+.GA;& ^|lv8|ܗ xϯ߼t\ET鬰H}Mq@5>(KM	wEw+Yї7Wmp﫬acPg9
,esN_o	8XrB-—kŇlGk^Vr{{PTm==(&O=ռ}#ֿIwSi~jV?^+m#YKdfD}x`.s
Zl-VS%{0}"ŴRI
MT*Z*Zj}t	_IZ˻fdQ#c4m/ò+S#XsZ-*g]鄗LܪȁVO_锞ֻ#.٣ޮI<潔~wݖ1KwdL/p1}M<]֘2VݷXo^Ch.![gV5?l!69#^=}oi*T:D$
PRk6BbgG5^j77_7Yu*3qhoGZnq`rZ+\>C}v][p(E?a5ōsH:)wK
l637o^`륕۔B
/gI,u]nXں?~4t/,궔%A)f
>J6F)=m͎c-/Aqy*痫"/V?˳(g"kAδכVD}km8inxJX9-"g$k?,#VG#Zz8
cQQA"e0H`(FJ$\0o݈>w
mތ|,eK::[ƾUgַngq_0xq$$ּ-:tZ[C5V=mX2+\ln@gԊPKZ,6fU|AE5Z(	t59YM{fGoggg/]՚)%]m\6|_7Ƅw[gukZogr5
i`[ۖ;7iRNyֱir0*ԻkAjWCY!?60a,&%ɧ՟j{D*fX5+ߌSyG*Ѧ[
PT,4P-BRcc氐CMQ7
=Ѓ'1q&JP*gNړ8/zm5ElF"U2QUDRuNJߎ%!4
4׬be	l'DBɺh$4* o
9*.P4ʔakD#SfѵN4^u+)= IORoUƶ*JFܓV%mkc[+=it4QsDVu`'舤oTUA9tf#T#v=b3=ٺ˰ja.#Nm}jPorq
3&GDU^*dT"Ɍ^AA$kl߃͇~9~~[+NK!
/)SJNNݿo3|2k3w
c('ٰpHQ);x2$”Oψ~;a|ȩ$[@_cvm"e9Sx|*aŢrÓm[9_b{.0xv6ii@݈|F0$
^,'&Þ^Foȁˤj|O-`fɩ9TaVQ^p&
!
BF&gKrƕ:=z>1.VtjY}{N	0
Ȗ>1YOT9*7yӮchO	gWy6R!2ݬ"lUxI8ɗbete2-'+)~pFVoޟ
Mfad#(fdta]\׏^yc9jUCڪE֚8W%`l[5UlU8$	`rU64]tT{m{1NUd/q&Tj,y_xkhH'_a]3_^}0RQ;j"X[doƐy筑Fȟ3w9
~ޑn7',[
ac;rל@ӓ,hGv)JJVHFOO.ڡjk bNT5+Qe!GGј>RkgjV 9KZ;>1h(em96e	ZML /$8YN&] _Y7@a$48b݀i{oAp~(^_*{TJ&\'UR\OMwVs^),.ZsuM3%6F.o y&hHuQM*WVڹy
5\9=O	njB S&UUJ*_\-5_e&#YG\/P0k
`A IQK"gOYbImt>險Ӊ!%m).@_KF
42zc8pe1sAL;#5t`H[j1x2X#CU*+7u $BI$?Ep76/dÃvfe1@m^)Ô=]E_ρrpRS+E6}K%aę>Ȏ,ڻ]./V
]x9Z;w=hbkR~8?xX}@KWU6qkPPq*vf
S4rj]F6,unzNFMy	ڍ(dE$:gҢ֬*RpN4Jۢ'嬜1_|yK${k69y$DG9zLZ
v}L4pHHS:YAԘAt#ˍÉKgG*$["SIPG*qXim/9l}3s	s{hԇq0Eh[o4aGiD7{vʙfU{YfGi;NEY	Uβ>L[ڇβ>X(Y	},G/::q.XJj:PqdՁt!tuaoI~; },ƙt?eS~wCOmAw~}$?)}ĠۄC'^Ax>Vzԉh>3;8Y`B㎌wBqA}Fvg̳%NC+
޵5#n$~q>SS'3I%yYO@Lt"KZ]r;4(YdEbE][?t7f#^wW(
@i4pZ(UGrY5?#0dw⻚\*ajZDn]Qay-)z-+dϻw%öf*ƫ?*ƫ=b
K;b
H(i1^Ha;SZW1^1f۫G1^	ۜS1^i(C=b
3F!b
sʷWb
$踊*h/긊*hs{{*ƫh{/Ee
*ƫ=0/	_RF1^X]1^h;&H0&z{+ƫ}qdxgJqTxaytxooqkxtWmm=զHy[cPORUʴ7q;~w^N1n/JiVo}n݂8A^#ޭ>
a?x
!~dݾbo'Mu۞MrZzf+m}DNHG]Mj+XP4
dFF4Ǹ󅶫zmLjmWx:c␣;!k}k&٤⺕"ka#wFZ?j*让hrY?O5B0fIb؛5qHl9i9z6ˆCԋԋg;\ɗӢvu߯sF3&j}!^dؚ%&ˍ39La=E+LF@U~z866;TQ;!aPZgɻnU`I6*?ϯ'}餐'	q'7tJ餓'Uq2a"s%:Mo*=O\&WN?LF>K#G0-}+3o64I1xE'XbnqB͹U-9jIO"\1yzg?dt¾c`w'[&݅imA#mA3"NNj =YH#g‰>*+L'i@'=?p5}W0P繒Ɯ?[#C.4'ݛp8a<	d4|0J^8AX/zW406=O/.*{ϭkK/$b$dD2QN\8BYFw{'S|Ť^$_^C5xCphp鍤)8xpp@$FKxL7XPp!jcB@h89-qlӯ>X	uAӛW|^B>mZ8ˁ:0Z8]s<IEMQ%wD!,E"0ӈ(/}|}7wuPmpJ3j%&Gzb'\[mYss	JgrqԊbl"ނBqq`R X3s̜j8xHH2nCdH_Qh	ǐ^OTtda`tٖgZ?Oߛˉ(m4
O@?]6zKcB:];CԜfy3~
c\Pʀs˟Ɯ0
[;@\Ȩ8J{П{
|LH/z;Д{Xe2_'DGW|=*M~z{C?t'~u*TІS2"-!CkP.}B.“ckJTM8_a{fE,zb [_LЉɻ8=c0#\V֛BdmF!%ۜ"F|K8HN,.#-0od<'	/'?T+8z6^yt))L5 Nuz1./1f	38gω2X)v+w\cNsu`FmWFFN|f6Y"šfLy5Dy&\rcJ('`y#{Qh,σ{  H`d1D]/;yxcX7<AlOS &ɏ^o߭I"$	C8t,vh0v_7+XKҧwZ	?hW܊x̢TFB[r[qUF$<c&yP<
Rp-o/$vafElh73N`ρf왷|䃟L9Dk!?8Ekq$0[MS/)ۡw]`NZ*mjy߻?û :uQŻs!e鬃m{ޭ	]o3,7-?'^	h9,/LQҨrcnv6At^QG*yp.h>j}eeu1Y/ɪi!hL<R)\65TfOuW佾7wAg%/߼ÌbuO*7.ߗIVk
Dl%w0R:x%(Bލ0VleS}(ㄢwᤆ
czM'I`FZ	242G[W X[?`;Fϡ^i,}Z('L<$FD~".CLpq'"UGJN;/0֠G`
A'Az[S`A'"R0oxjq5n,Aqrca!u[ìF*).J1bUB.eTe;Ze$M0M:Qh,od"rm2n {ʽ3o
t9'1ƐRdaϑHݵ\ePIxed,q`+kLɩ\2YĜ繱2Hz&yE \ꂹ"	nJBGѵa(4$FAgIGTJˍk!9!K]E.gl0DZ	=ZmpUdk\?P-W{=Q(Fq
G 4Eq[H[qL[͉{#mL
)]wS4bZLD&$$PQp%(gr- κ[*ѦXp!wj
*v*E"sVdU:[͏u"qJzA6ьh@{X[sDUrZ\;WVfžb;\>@h(7֩\|c@5ou[UwGB5%Ebבo~
eEm𗝮P6[RBwwj]R;JSX5h|ݠqvm^4V(x
_ձR{5nm.5xZ>_ޓZVrr]b*!/6-m\?ێ|
	ũMz0nVbCϨHMJGn+>"[)ZS-C8{7AV]Sgһծn
0Ҽ'jӭj
ݫ1n
o6շo L*k{ZsufM)E.@].$;a'$D~IJ0`Ȳym}~>pV0x[$tt2N:ҏF+٬5K5dbECHX)G>C.Z oeSdJouySw߆쨓f;o-d掚K-.lܪ8YԒEWsa7]`t)"c:ᔝ_UAOi-
C8~W+'T-JKuPqaD&|TLVc0(vy9Ǥ5#唡	\zmYss!ɹva/ĝ޹[rLֻQoHt{Pm:6oEkިM7B\エV_0ͦwYeXKqE2tu]*-^W\fa]-rd/c?ߠ--'r`#~,?t[7{M%{ש 8J

0 KQH>UPZ?BTnXLM5`Z/θ p3rA\Qf1InjS
ͤ
vԗ&Lv،űΒ0y,e,3'>krR.
q,}L{Ry'>?b0eKuESv,}\Urbf)q,%4NA1/I)?oRR*6TKrUj):Y82KrUj=#p]D޵5˩$g(4WWaVRƮzr`0cJT4sP&qx߇FKXԗe.t9Y
,h%{R`a,ͥVfiP9k^0ҥ7OװRa,e/d)0RsN+q0rG`/N,N,=btsxk[ԗ%ǞXz,$0i/XJIKRk8f)0"wO{Rda,Ʉ>7KYکU/$5%Xz,eTꯤBfiP9K4Z;Ҡ>syLQ`)Ua,]JN'QGRFX2,e$l'5Ky_ʽJC/X\jO~QJ
Gv5Fk~x:Yƣ̚flUuss7[3&ӻ̡5f	uz3#,BV,%	8G|7(w^hUJ|)-s
r:;uICN"*TAݟ -ߐ,;5"Sc`B
JNm濳,=B{֗9Q'7}Y_aњc^^p._m&Mv}#clR3S'k@*b*R,5R)J/M87tRca+QoJ5 /ION)l.'aґӇw?<_ҹ_iT~?=ĵ`Z&7 -atZۃ
DyGR/%O
JOAx%dF9Yb
GsoSO<|-}!,7(Sɜ[d@6Jin%Be
VVA!qv8=
RsuOku[Iu[ňRg	Qh2P'Һh&Ke	G%h4Eʬ&f1UC.滷L-)Z޺j>m`W_kWlDbT*P;vgM_l<}wA|ux',Z1+
zTK=ﷷ??|/F?1,Qk<^' "m#w"Α.S>hYDstۈ6q[S-l3
Y&yWVքXn~P(TorKE`rpȴ$H_A;j$>vz=>3{ƺy9A6]7U8~(Iß_w~&Js!,۩э\QBED *ޤ~y{t6_!a1;w+87q#y6dz_WP+K')Z@r6[){ض9$}5vͤ>Gn=/Wﶗ,pVbds)36ݚ?
8[,*\.6ϯ7We9{0=sDCFF^!QBSڐfF
JA.IΟA߮^EIyAn{#]AqmkN)!GpNclIZXg~C
pgOfGoYo v|BRgUF	ܸ	g\}xOJJ ׽\TBE(}Wgd>ԁ/_51
vwU\qv8!BcK1;]G`6n1cٳ,+	MLpi©Y
>(=RG`sC`CzAXKc3b߇Y^nfh4HNm6вT33#s9`	琐ݷzT֯jrܰcσ]xJKUyy)eVэ}!-?FePyk<:3%3*ވoS7/)u,i^#
W0a	WBE5owŶsRQa|U**5J9hf%D5(&Tu:'
;1**A*A⺐u~B}γ**KOh8Ըzq4_B~IOu(~jԚpU$!:xJIg\LĄ17$3	͜z`N6Iu}+ݳ b턓愁G@Jpy
7N6u
bNj‡#3RB!Q+˔FN{
܏/HoB.RKCX]Fp-S:5WNZRkus/uFv@Ke\vfZ&)ZSn:87V :uQźrH^֭~5@C^9EK
F
i5?a*nYvy8^>|N,Iv~k?FWQ;1u|}vE%kO8j}42<>$ϥb~q77uv=OޜCo?_Ɲznx<`~3rRNmyӟ7ˢ.~QCX^ ~~{l9nܪyPuvD2bQDXQX7o}J؟l1/Chֱ@,c;?:)Y#'kq
Wt{&u!G$pyR}[ Iv4	'FN(=$(wEd-<?~g͡F)ݦg]jhWkR#.;ފD*2-4c"4q4fɳ674W)
@hBTnKG
f݁RN1v,<%8~(_ܱzq;r1151nEW݃1$'eUC"m	sU
!UO*LP͟_niچwP%fi= hih
Hf*,R`$c!HNP	y;X9D\Ý!|W	7sA@KTsx4̈́>_av0I{VEW'%'[UW?u)?I)lg{K0WE?;_R:LA}gq.EZ^!?EXKԊ0:%HF@c"IQTQ@hg	5jyUUURT[/FgUk]
7zdLƚ:3-QL[LP[i,N<:i--f48
Y67xa5I25&qJ9Xc5R0Z1%164`Kuvu&XWuhĔGÍHKy*VފYڒLg"M8IIj4T[Ƴlc6rS#!g`d54uO,DA4JPHHv<4=@Sfcmm֡,^2:o;a5XiP_&) \%Ib!Iq:j+"B:D˰,S
JJ27kin9Sʢ=WI!٭$'\ cZR()*4
B 5AІO=tL@Sa iqg":ZjLJKce}+qDV؈RbXei[c@ocI	h<(+	G/55,G@[f[30FiV?	2BkuAk
S@&-h!Pn;X9ItG@K1>u60`+>ec\=t;Fc='
a+Q?{0Rp6g
-C{FN)xf0e+e+Xɵ[SV,[_m
3$mekogeA#W՛dY=e*$O
Ю
b"ٲe!ɗ,+:̓GSTo{BcQT;笃fK]vH?=0_MѬҫ?g?^_7]|=#x6_=c-eՀ
wky=p`d˂H
7G`?k;-H!5
9^1peeJ#dL1#9;?X=B7hS^KBUh8v;!>܌rlgAu9/G5o|9g@`
pWH`RP%Ų29F\h!4Y!%Ze<EGHmPF)@\{˝XX_O/ռ#&ںKZ>?}yiV3yPE^ùꈊ蛳VݤW.T;;)R8zxSmJ/WyrMK][ulLS`-m^\\/Y_/6rkzywzL9x||C.}
R
Snhn%E8EBUܕϕ=dOUE|dPrHiBp43Yibqf	[,vmo~YN2Y4c9l݇,_+W%__-^_u~该5_w9I?TPrI?R)f}`f_!‹<	?'_sDc8Yu4[+MU`ʝuu.ڋZAWc/ԶVB39stؠP
ʱ)QzHESjZlBT1ϕ4˵CaM5QK_j(PzHECjW!'5JC(+cGCR/R4fBY<#>i#gwRs{:ogGef(Eψ_J
Pz~(k#{k)PzXESjF~(nYtv3JCi)RjD7J%TnmGCR/R#1L(=GC)hڇEVԋԨPz(E˪X` ^<ڪi]z((u鴰+-Һ_Ғ]BKbrf)gy^VdI̤K[fAnTV$`2,Mi|ߕPHd=^iy|ܿח׿U{/=u%Z#4H_Ա0ZR+i9l\ݼ._:ns1G*߭y[TGc\91D3;N(NQ 3o:
9ĕ'J՜kmX!1ؽ|B5ySYlDLA6%U$F4HɅI2j`@e,Om(.
ÑFgX2@@
R<,} Ѩ%ʙ|ly/>_P#V{|X=X|.|'wXy;J˧7~|7r)FK['!;$!;\BԵF*0zNaH_tʑښOESj-]Y8=g1mP@ַq"N/fv~OoV޹ZtWj͘ľ
&+?"|w?'ljvܮZWI}vP!w^==fʊŠ'|01ػ}	[|vEHiˡ[9w!9DcJI`\:fn[|w"CM]QߡDYTKyq39zA~gO[RMd?}*WUeFRԨT|v(?)var/home/core/zuul-output/logs/kubelet.log0000644000000000000000003777231715134375476017735 0ustar  rootrootJan 22 09:42:46 crc systemd[1]: Starting Kubernetes Kubelet...
Jan 22 09:42:46 crc restorecon[4743]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:46 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0
Jan 22 09:42:47 crc restorecon[4743]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0
Jan 22 09:42:47 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 22 09:42:47 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version.
Jan 22 09:42:47 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 22 09:42:47 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 22 09:42:47 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jan 22 09:42:47 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.234550    4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.237988    4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238013    4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238017    4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238021    4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238026    4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238030    4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238034    4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238037    4824 feature_gate.go:330] unrecognized feature gate: NewOLM
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238041    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238045    4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238048    4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238052    4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238057    4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238062    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238066    4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238071    4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238076    4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238081    4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238086    4824 feature_gate.go:330] unrecognized feature gate: SignatureStores
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238091    4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238095    4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238100    4824 feature_gate.go:330] unrecognized feature gate: OVNObservability
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238105    4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238111    4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238117    4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238122    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238126    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238130    4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238134    4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238137    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238141    4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238151    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238156    4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238160    4824 feature_gate.go:330] unrecognized feature gate: PinnedImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238164    4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238169    4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238174    4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238178    4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238182    4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238186    4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238191    4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238195    4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238200    4824 feature_gate.go:330] unrecognized feature gate: Example
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238204    4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238207    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238211    4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238215    4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238219    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238222    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238226    4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238229    4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238233    4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238237    4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238241    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238244    4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238247    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238251    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238254    4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238257    4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238261    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238264    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238269    4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238272    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238276    4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238280    4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238283    4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238286    4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238290    4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238293    4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238297    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.238300    4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238391    4824 flags.go:64] FLAG: --address="0.0.0.0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238400    4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238408    4824 flags.go:64] FLAG: --anonymous-auth="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238415    4824 flags.go:64] FLAG: --application-metrics-count-limit="100"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238420    4824 flags.go:64] FLAG: --authentication-token-webhook="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238424    4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238430    4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238436    4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238440    4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238444    4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238449    4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238453    4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238458    4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238464    4824 flags.go:64] FLAG: --cgroup-root=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238469    4824 flags.go:64] FLAG: --cgroups-per-qos="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238473    4824 flags.go:64] FLAG: --client-ca-file=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238477    4824 flags.go:64] FLAG: --cloud-config=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238481    4824 flags.go:64] FLAG: --cloud-provider=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238485    4824 flags.go:64] FLAG: --cluster-dns="[]"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238490    4824 flags.go:64] FLAG: --cluster-domain=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238494    4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238498    4824 flags.go:64] FLAG: --config-dir=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238502    4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238507    4824 flags.go:64] FLAG: --container-log-max-files="5"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238512    4824 flags.go:64] FLAG: --container-log-max-size="10Mi"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238516    4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238521    4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238525    4824 flags.go:64] FLAG: --containerd-namespace="k8s.io"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238529    4824 flags.go:64] FLAG: --contention-profiling="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238534    4824 flags.go:64] FLAG: --cpu-cfs-quota="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238538    4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238542    4824 flags.go:64] FLAG: --cpu-manager-policy="none"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238546    4824 flags.go:64] FLAG: --cpu-manager-policy-options=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238551    4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238555    4824 flags.go:64] FLAG: --enable-controller-attach-detach="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238559    4824 flags.go:64] FLAG: --enable-debugging-handlers="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238563    4824 flags.go:64] FLAG: --enable-load-reader="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238567    4824 flags.go:64] FLAG: --enable-server="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238571    4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238577    4824 flags.go:64] FLAG: --event-burst="100"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238581    4824 flags.go:64] FLAG: --event-qps="50"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238585    4824 flags.go:64] FLAG: --event-storage-age-limit="default=0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238590    4824 flags.go:64] FLAG: --event-storage-event-limit="default=0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238594    4824 flags.go:64] FLAG: --eviction-hard=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238600    4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238605    4824 flags.go:64] FLAG: --eviction-minimum-reclaim=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238609    4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238613    4824 flags.go:64] FLAG: --eviction-soft=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238617    4824 flags.go:64] FLAG: --eviction-soft-grace-period=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238621    4824 flags.go:64] FLAG: --exit-on-lock-contention="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238626    4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238630    4824 flags.go:64] FLAG: --experimental-mounter-path=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238634    4824 flags.go:64] FLAG: --fail-cgroupv1="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238638    4824 flags.go:64] FLAG: --fail-swap-on="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238642    4824 flags.go:64] FLAG: --feature-gates=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238647    4824 flags.go:64] FLAG: --file-check-frequency="20s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238651    4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238655    4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238659    4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238663    4824 flags.go:64] FLAG: --healthz-port="10248"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238667    4824 flags.go:64] FLAG: --help="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238671    4824 flags.go:64] FLAG: --hostname-override=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238675    4824 flags.go:64] FLAG: --housekeeping-interval="10s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238679    4824 flags.go:64] FLAG: --http-check-frequency="20s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238683    4824 flags.go:64] FLAG: --image-credential-provider-bin-dir=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238687    4824 flags.go:64] FLAG: --image-credential-provider-config=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238691    4824 flags.go:64] FLAG: --image-gc-high-threshold="85"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238695    4824 flags.go:64] FLAG: --image-gc-low-threshold="80"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238699    4824 flags.go:64] FLAG: --image-service-endpoint=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238703    4824 flags.go:64] FLAG: --kernel-memcg-notification="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238707    4824 flags.go:64] FLAG: --kube-api-burst="100"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238711    4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238715    4824 flags.go:64] FLAG: --kube-api-qps="50"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238719    4824 flags.go:64] FLAG: --kube-reserved=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238723    4824 flags.go:64] FLAG: --kube-reserved-cgroup=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238727    4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238733    4824 flags.go:64] FLAG: --kubelet-cgroups=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238737    4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238741    4824 flags.go:64] FLAG: --lock-file=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238745    4824 flags.go:64] FLAG: --log-cadvisor-usage="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238750    4824 flags.go:64] FLAG: --log-flush-frequency="5s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238754    4824 flags.go:64] FLAG: --log-json-info-buffer-size="0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238760    4824 flags.go:64] FLAG: --log-json-split-stream="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238764    4824 flags.go:64] FLAG: --log-text-info-buffer-size="0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238768    4824 flags.go:64] FLAG: --log-text-split-stream="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238772    4824 flags.go:64] FLAG: --logging-format="text"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238777    4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238781    4824 flags.go:64] FLAG: --make-iptables-util-chains="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238785    4824 flags.go:64] FLAG: --manifest-url=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238790    4824 flags.go:64] FLAG: --manifest-url-header=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238798    4824 flags.go:64] FLAG: --max-housekeeping-interval="15s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238802    4824 flags.go:64] FLAG: --max-open-files="1000000"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238807    4824 flags.go:64] FLAG: --max-pods="110"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238812    4824 flags.go:64] FLAG: --maximum-dead-containers="-1"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238816    4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238820    4824 flags.go:64] FLAG: --memory-manager-policy="None"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238824    4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238828    4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238832    4824 flags.go:64] FLAG: --node-ip="192.168.126.11"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238836    4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238846    4824 flags.go:64] FLAG: --node-status-max-images="50"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238850    4824 flags.go:64] FLAG: --node-status-update-frequency="10s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238854    4824 flags.go:64] FLAG: --oom-score-adj="-999"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238858    4824 flags.go:64] FLAG: --pod-cidr=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238862    4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238868    4824 flags.go:64] FLAG: --pod-manifest-path=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238872    4824 flags.go:64] FLAG: --pod-max-pids="-1"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238876    4824 flags.go:64] FLAG: --pods-per-core="0"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238880    4824 flags.go:64] FLAG: --port="10250"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238885    4824 flags.go:64] FLAG: --protect-kernel-defaults="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238889    4824 flags.go:64] FLAG: --provider-id=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238893    4824 flags.go:64] FLAG: --qos-reserved=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238897    4824 flags.go:64] FLAG: --read-only-port="10255"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238901    4824 flags.go:64] FLAG: --register-node="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238905    4824 flags.go:64] FLAG: --register-schedulable="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238910    4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238917    4824 flags.go:64] FLAG: --registry-burst="10"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238921    4824 flags.go:64] FLAG: --registry-qps="5"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238925    4824 flags.go:64] FLAG: --reserved-cpus=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238929    4824 flags.go:64] FLAG: --reserved-memory=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238934    4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238938    4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238945    4824 flags.go:64] FLAG: --rotate-certificates="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238950    4824 flags.go:64] FLAG: --rotate-server-certificates="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238954    4824 flags.go:64] FLAG: --runonce="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238958    4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238962    4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238966    4824 flags.go:64] FLAG: --seccomp-default="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238970    4824 flags.go:64] FLAG: --serialize-image-pulls="true"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238974    4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238978    4824 flags.go:64] FLAG: --storage-driver-db="cadvisor"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238982    4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238986    4824 flags.go:64] FLAG: --storage-driver-password="root"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238990    4824 flags.go:64] FLAG: --storage-driver-secure="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238994    4824 flags.go:64] FLAG: --storage-driver-table="stats"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.238998    4824 flags.go:64] FLAG: --storage-driver-user="root"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239003    4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239007    4824 flags.go:64] FLAG: --sync-frequency="1m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239011    4824 flags.go:64] FLAG: --system-cgroups=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239015    4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239021    4824 flags.go:64] FLAG: --system-reserved-cgroup=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239028    4824 flags.go:64] FLAG: --tls-cert-file=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239032    4824 flags.go:64] FLAG: --tls-cipher-suites="[]"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239037    4824 flags.go:64] FLAG: --tls-min-version=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239041    4824 flags.go:64] FLAG: --tls-private-key-file=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239045    4824 flags.go:64] FLAG: --topology-manager-policy="none"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239049    4824 flags.go:64] FLAG: --topology-manager-policy-options=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239053    4824 flags.go:64] FLAG: --topology-manager-scope="container"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239057    4824 flags.go:64] FLAG: --v="2"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239063    4824 flags.go:64] FLAG: --version="false"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239069    4824 flags.go:64] FLAG: --vmodule=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239074    4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239079    4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239180    4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239191    4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239195    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239199    4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239203    4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239207    4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239210    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239214    4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239218    4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239221    4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239225    4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239229    4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239232    4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239236    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239239    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239243    4824 feature_gate.go:330] unrecognized feature gate: NewOLM
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239247    4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239250    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239254    4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239257    4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239263    4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239267    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239270    4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239274    4824 feature_gate.go:330] unrecognized feature gate: OVNObservability
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239278    4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239282    4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239288    4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239292    4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239297    4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239301    4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239304    4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239308    4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239313    4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239318    4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239323    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239327    4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239332    4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239336    4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239340    4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239343    4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239347    4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239367    4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239371    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239375    4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239379    4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239382    4824 feature_gate.go:330] unrecognized feature gate: SignatureStores
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239386    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239390    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239394    4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239398    4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239402    4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239406    4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239412    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239415    4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239419    4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239422    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239426    4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239429    4824 feature_gate.go:330] unrecognized feature gate: Example
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239433    4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239436    4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239440    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239443    4824 feature_gate.go:330] unrecognized feature gate: PinnedImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239447    4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239450    4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239454    4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239458    4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239461    4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239465    4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239468    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239471    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.239475    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.239487    4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]}
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.249663    4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.249706    4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249809    4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249819    4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249826    4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249850    4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249856    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249861    4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249866    4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249872    4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249876    4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249881    4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249885    4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249889    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249894    4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249898    4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249902    4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249907    4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249911    4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249916    4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249920    4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249926    4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249932    4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249940    4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249946    4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249951    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249956    4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249963    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249968    4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249972    4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249977    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249983    4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249987    4824 feature_gate.go:330] unrecognized feature gate: Example
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249992    4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.249997    4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250002    4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250008    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250013    4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250017    4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250021    4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250026    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250030    4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250035    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250039    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250046    4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250051    4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250056    4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250061    4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250066    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250070    4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250077    4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250082    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250087    4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250092    4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250096    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250101    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250105    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250110    4824 feature_gate.go:330] unrecognized feature gate: NewOLM
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250115    4824 feature_gate.go:330] unrecognized feature gate: SignatureStores
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250119    4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250123    4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250127    4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250132    4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250136    4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250140    4824 feature_gate.go:330] unrecognized feature gate: OVNObservability
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250144    4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250149    4824 feature_gate.go:330] unrecognized feature gate: PinnedImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250153    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250157    4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250161    4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250165    4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250170    4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250178    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.250187    4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]}
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250343    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250350    4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250372    4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250376    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250382    4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250388    4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250393    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250398    4824 feature_gate.go:330] unrecognized feature gate: SignatureStores
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250404    4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250410    4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250415    4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250419    4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250424    4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250430    4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250436    4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250442    4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250448    4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250454    4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250459    4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250465    4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250469    4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250474    4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250479    4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250483    4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250488    4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250492    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250496    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250501    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250506    4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250511    4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release.
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250517    4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250522    4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250527    4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250532    4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250537    4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250542    4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250546    4824 feature_gate.go:330] unrecognized feature gate: Example
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250550    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250554    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250559    4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250563    4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250568    4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250572    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250576    4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250581    4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250585    4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250591    4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250595    4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250600    4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250605    4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250609    4824 feature_gate.go:330] unrecognized feature gate: PinnedImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250614    4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250619    4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250623    4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250628    4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250632    4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250636    4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250641    4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250645    4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250650    4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250654    4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250658    4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250663    4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250667    4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250672    4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250676    4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250681    4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250686    4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250691    4824 feature_gate.go:330] unrecognized feature gate: OVNObservability
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250695    4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.250700    4824 feature_gate.go:330] unrecognized feature gate: NewOLM
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.250709    4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]}
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.251141    4824 server.go:940] "Client rotation is on, will bootstrap in background"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.254415    4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.254512    4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem".
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.255063    4824 server.go:997] "Starting client certificate rotation"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.255085    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.255270    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-09 04:26:46.637856141 +0000 UTC
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.255421    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.265877    4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.267435    4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.268555    4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.279959    4824 log.go:25] "Validated CRI v1 runtime API"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.292756    4824 log.go:25] "Validated CRI v1 image API"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.294328    4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.297073    4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-22-09-37-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3]
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.297124    4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}]
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.324774    4824 manager.go:217] Machine: {Timestamp:2026-01-22 09:42:47.322881562 +0000 UTC m=+0.226358585 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654116352 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:2a2d1f9c-78d4-4a51-9bdc-24b710991b2b BootID:b28d29c1-bb01-432c-b92a-540ec89f6524 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108168 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827056128 Type:vfs Inodes:4108168 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:50:b2:a5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:50:b2:a5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c1:83:29 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a4:5c:6d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:7f:85:e6 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ab:ee:fd Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:2e:95:d5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:02:f4:64:1d:82:e5 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:22:33:16:81:36:2d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654116352 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None}
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.325134    4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available.
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.325316    4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:}
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.326313    4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.326544    4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.326595    4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2}
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.326858    4824 topology_manager.go:138] "Creating topology manager with none policy"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.326872    4824 container_manager_linux.go:303] "Creating device plugin manager"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.327069    4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.327114    4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.327413    4824 state_mem.go:36] "Initialized new in-memory state store"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.327503    4824 server.go:1245] "Using root directory" path="/var/lib/kubelet"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.328174    4824 kubelet.go:418] "Attempting to sync node with API server"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.328199    4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.328228    4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.328247    4824 kubelet.go:324] "Adding apiserver pod source"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.328260    4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.330578    4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.330811    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.330935    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.330813    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.330999    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.331098    4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem".
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332105    4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332743    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332774    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332784    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332794    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332811    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332820    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332828    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332843    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332853    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332862    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332876    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.332885    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.333593    4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.334344    4824 server.go:1280] "Started kubelet"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.334409    4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.335311    4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.335409    4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Jan 22 09:42:47 crc systemd[1]: Started Kubernetes Kubelet.
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.338698    4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.338527    4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188d044f50fcd2a4  default    0 0001-01-01 00:00:00 +0000 UTC   map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:42:47.334146724 +0000 UTC m=+0.237623727,LastTimestamp:2026-01-22 09:42:47.334146724 +0000 UTC m=+0.237623727,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.341627    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.341680    4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.344980    4824 server.go:460] "Adding debug handlers to kubelet server"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.345066    4824 volume_manager.go:287] "The desired_state_of_world populator starts"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.345088    4824 volume_manager.go:289] "Starting Kubelet Volume Manager"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.345265    4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.345317    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 08:10:18.099761806 +0000 UTC
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.345462    4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.346180    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.346256    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.347938    4824 factory.go:55] Registering systemd factory
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.347986    4824 factory.go:221] Registration of the systemd container factory successfully
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.348070    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="200ms"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.348329    4824 factory.go:153] Registering CRI-O factory
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.348374    4824 factory.go:221] Registration of the crio container factory successfully
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.348456    4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.348484    4824 factory.go:103] Registering Raw factory
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.348500    4824 manager.go:1196] Started watching for new ooms in manager
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.349372    4824 manager.go:319] Starting recovery of all containers
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.355700    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.355998    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356143    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356254    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356378    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356484    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356590    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356717    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.356875    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357021    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357140    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357261    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357408    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357536    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357650    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357756    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357863    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.357961    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.358089    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.358202    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.358291    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.358418    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.358518    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.359379    4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.359508    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.359875    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360221    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360389    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360615    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360722    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360749    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360774    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360807    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360830    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360845    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360869    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.360887    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361133    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361150    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361166    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361189    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361207    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361228    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361242    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361260    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361283    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361297    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361320    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361400    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361424    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361446    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361460    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361480    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361503    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361530    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361557    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361574    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361597    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361612    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361634    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361648    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361660    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361711    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361725    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361744    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361760    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361776    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361794    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361809    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361832    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361847    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361865    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361888    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361907    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361933    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361952    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361968    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.361995    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362013    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362042    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362063    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362080    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362111    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362130    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362154    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362170    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362191    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362740    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362760    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362778    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362806    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362822    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362846    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362863    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362878    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362899    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362916    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362938    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362955    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362973    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.362997    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363012    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363033    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363113    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363135    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363172    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363249    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363277    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363299    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363328    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363386    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363407    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363433    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363451    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363473    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363491    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363517    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363534    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363555    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363572    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363588    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363613    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363629    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363653    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363667    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363682    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363702    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363715    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363736    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363750    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363763    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363784    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363798    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363816    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363841    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363856    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363881    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363898    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363914    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363940    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363955    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.363989    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364006    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364022    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364043    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364060    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364085    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364101    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364116    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364145    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364160    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364188    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364204    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364229    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364258    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364274    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364294    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364309    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364326    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364376    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364393    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364411    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364431    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364447    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364474    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364492    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364518    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364541    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364558    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364581    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364597    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364613    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364639    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364667    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364686    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364703    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364717    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364735    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364749    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364768    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364782    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364795    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364813    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364827    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364847    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364866    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364881    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364908    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364944    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364959    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364981    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.364994    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365022    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365039    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365062    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365086    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365107    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365140    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365158    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365172    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365191    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365207    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365227    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365243    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365259    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365281    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365296    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365317    4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext=""
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365331    4824 reconstruct.go:97] "Volume reconstruction finished"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.365341    4824 reconciler.go:26] "Reconciler: start to sync state"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.372316    4824 manager.go:324] Recovery completed
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.385848    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.387605    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.387655    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.387664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.388572    4824 cpu_manager.go:225] "Starting CPU manager" policy="none"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.388590    4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.388614    4824 state_mem.go:36] "Initialized new in-memory state store"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.398086    4824 policy_none.go:49] "None policy: Start"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.399622    4824 memory_manager.go:170] "Starting memorymanager" policy="None"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.399663    4824 state_mem.go:35] "Initializing new in-memory state store"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.400220    4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.402779    4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.402846    4824 status_manager.go:217] "Starting to sync pod status with apiserver"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.402882    4824 kubelet.go:2335] "Starting kubelet main sync loop"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.402930    4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.403604    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.403683    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.445992    4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.469400    4824 manager.go:334] "Starting Device Plugin manager"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.469457    4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.469472    4824 server.go:79] "Starting device plugin registration server"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.469969    4824 eviction_manager.go:189] "Eviction manager: starting control loop"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.469990    4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.470298    4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.470516    4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.470548    4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.476339    4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.503970    4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"]
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.504055    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.505167    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.505230    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.505241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.505451    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.505712    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.505756    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506516    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506566    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506692    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506809    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506825    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506846    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.506892    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.507766    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.507808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.507829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.507842    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.507832    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.507905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.508044    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.508112    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.508140    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509048    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509073    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509085    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509089    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509132    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509332    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509509    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.509559    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510216    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510253    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510266    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510475    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510534    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510489    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510580    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.510544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.511378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.511421    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.511437    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.549123    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="400ms"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568471    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568520    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568547    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568573    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568707    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568747    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568775    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568863    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568902    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568921    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568937    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568951    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568966    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.568992    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.569011    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.570513    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.575796    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.575849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.575867    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.575913    4824 kubelet_node_status.go:76] "Attempting to register node" node="crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.576896    4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670106    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670164    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670197    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670223    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670252    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670274    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670289    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670335    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670326    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670299    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670424    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670486    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670403    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670521    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670538    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670534    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670388    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670554    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670560    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670574    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670578    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670550    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670672    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670730    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670728    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670768    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670784    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670776    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670859    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.670946    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.776978    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.778061    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.778095    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.778106    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.778137    4824 kubelet_node_status.go:76] "Attempting to register node" node="crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.778558    4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.845204    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.866145    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-95233f4420869c992ec32e23ff45a0a8a6743fd6279d8121b3b00546628a907a WatchSource:0}: Error finding container 95233f4420869c992ec32e23ff45a0a8a6743fd6279d8121b3b00546628a907a: Status 404 returned error can't find the container with id 95233f4420869c992ec32e23ff45a0a8a6743fd6279d8121b3b00546628a907a
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.871830    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.888243    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-cbd6fa43f50c1080ec12cf1f5cda4f488877d3fe6c886c292f075c5e5600d911 WatchSource:0}: Error finding container cbd6fa43f50c1080ec12cf1f5cda4f488877d3fe6c886c292f075c5e5600d911: Status 404 returned error can't find the container with id cbd6fa43f50c1080ec12cf1f5cda4f488877d3fe6c886c292f075c5e5600d911
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.910723    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.924652    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7e92a0486fd0adde674521dc4bc790c98356db89e61bbdb92162a0669eea103a WatchSource:0}: Error finding container 7e92a0486fd0adde674521dc4bc790c98356db89e61bbdb92162a0669eea103a: Status 404 returned error can't find the container with id 7e92a0486fd0adde674521dc4bc790c98356db89e61bbdb92162a0669eea103a
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.931227    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: I0122 09:42:47.937039    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.947387    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-7af79edda3c8f5b82771b9209dbfecc724fb611351490374d66e5e72e996f5dc WatchSource:0}: Error finding container 7af79edda3c8f5b82771b9209dbfecc724fb611351490374d66e5e72e996f5dc: Status 404 returned error can't find the container with id 7af79edda3c8f5b82771b9209dbfecc724fb611351490374d66e5e72e996f5dc
Jan 22 09:42:47 crc kubenswrapper[4824]: W0122 09:42:47.949412    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6ced4566b934beba173818a86e10d72779caa0a3a692318abe1d599cef0e6036 WatchSource:0}: Error finding container 6ced4566b934beba173818a86e10d72779caa0a3a692318abe1d599cef0e6036: Status 404 returned error can't find the container with id 6ced4566b934beba173818a86e10d72779caa0a3a692318abe1d599cef0e6036
Jan 22 09:42:47 crc kubenswrapper[4824]: E0122 09:42:47.949624    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="800ms"
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.164673    4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188d044f50fcd2a4  default    0 0001-01-01 00:00:00 +0000 UTC   map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:42:47.334146724 +0000 UTC m=+0.237623727,LastTimestamp:2026-01-22 09:42:47.334146724 +0000 UTC m=+0.237623727,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.179183    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.180648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.180691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.180701    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.180875    4824 kubelet_node_status.go:76] "Attempting to register node" node="crc"
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.181858    4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc"
Jan 22 09:42:48 crc kubenswrapper[4824]: W0122 09:42:48.267933    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.268428    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:48 crc kubenswrapper[4824]: W0122 09:42:48.317835    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.317999    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.335609    4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.345868    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 00:22:53.412466471 +0000 UTC
Jan 22 09:42:48 crc kubenswrapper[4824]: W0122 09:42:48.351917    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.351981    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.409547    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"95233f4420869c992ec32e23ff45a0a8a6743fd6279d8121b3b00546628a907a"}
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.410978    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6ced4566b934beba173818a86e10d72779caa0a3a692318abe1d599cef0e6036"}
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.412511    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7af79edda3c8f5b82771b9209dbfecc724fb611351490374d66e5e72e996f5dc"}
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.414043    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7e92a0486fd0adde674521dc4bc790c98356db89e61bbdb92162a0669eea103a"}
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.415786    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cbd6fa43f50c1080ec12cf1f5cda4f488877d3fe6c886c292f075c5e5600d911"}
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.756931    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="1.6s"
Jan 22 09:42:48 crc kubenswrapper[4824]: W0122 09:42:48.757179    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.757264    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.982030    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.983938    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.984029    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.984041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:48 crc kubenswrapper[4824]: I0122 09:42:48.984085    4824 kubelet_node_status.go:76] "Attempting to register node" node="crc"
Jan 22 09:42:48 crc kubenswrapper[4824]: E0122 09:42:48.984699    4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.335202    4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.346853    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 16:33:01.568539975 +0000 UTC
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.421638    4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f" exitCode=0
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.421738    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.421900    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.423343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.423430    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.423449    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.426939    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.426979    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.426992    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.427009    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.427067    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.428456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.428521    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.428545    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.430090    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069" exitCode=0
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.430155    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.430245    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.431428    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.431468    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.431481    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.432541    4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb" exitCode=0
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.432631    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.432669    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.433665    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.433726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.433753    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.434572    4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048" exitCode=0
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.434615    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048"}
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.434613    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.434749    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:49 crc kubenswrapper[4824]: E0122 09:42:49.435959    4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.436007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.436040    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.436057    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.437496    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.438506    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.438548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:49 crc kubenswrapper[4824]: I0122 09:42:49.438571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.347575    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 12:44:56.412679001 +0000 UTC
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.440004    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.440058    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.441607    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.441636    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.441644    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.442665    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.442699    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.442710    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.442772    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.443686    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.443709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.443717    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.450111    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.450219    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.450238    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.450251    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.450278    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.450303    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.451035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.451055    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.451063    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.453631    4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7" exitCode=0
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.453678    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7"}
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.453743    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.453823    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.458125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.458168    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.458179    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.459050    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.459079    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.459090    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.585203    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.586393    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.586426    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.586435    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:50 crc kubenswrapper[4824]: I0122 09:42:50.586459    4824 kubelet_node_status.go:76] "Attempting to register node" node="crc"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.348478    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 01:40:33.643722272 +0000 UTC
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.459864    4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f" exitCode=0
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460001    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460033    4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460073    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460087    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f"}
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460136    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460005    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.460425    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461596    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461607    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461697    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461779    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.461982    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.462000    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.462250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.462278    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:51 crc kubenswrapper[4824]: I0122 09:42:51.462296    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.348681    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 04:43:02.812841815 +0000 UTC
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.468475    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.469556    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b0a0eec4fb2313dfd42a94f3a70cc7a509038dd5e624597cb05961d1dc1e3849"}
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.469627    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37af36f712c0370669b5316d98b277194b465ae3bd9197482859245a049ad0e7"}
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.469652    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ce73eb24a4e7c75b937de8520c772a2b37aa4e1f4aa142438701a243d1d2e3a5"}
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.469670    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"baa91323c49714652c26af8de91efa2ccbc2025ff4b7197e48bfa4aed0792a98"}
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.470763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.470817    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:52 crc kubenswrapper[4824]: I0122 09:42:52.470829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.347246    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.348963    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 22:10:12.470847382 +0000 UTC
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.474173    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"044fa372f704e0ec5d127058e2c208fbb15d0c7d1cba8497a9d495573e9ecdbf"}
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.474313    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.474651    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.475498    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.475564    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.475578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.476114    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.476164    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.476182    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:53 crc kubenswrapper[4824]: I0122 09:42:53.834377    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.280892    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.349270    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 08:57:22.690451506 +0000 UTC
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.476475    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.476476    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.477879    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.477945    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.477962    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.478406    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.478482    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:54 crc kubenswrapper[4824]: I0122 09:42:54.478588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.349958    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 22:36:55.860183342 +0000 UTC
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.499640    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.499882    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.501468    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.501525    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.501538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.779737    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:55 crc kubenswrapper[4824]: I0122 09:42:55.785542    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.003276    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.003608    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.005110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.005177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.005186    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.350595    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 22:05:45.842142681 +0000 UTC
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.481652    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.482829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.482899    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:56 crc kubenswrapper[4824]: I0122 09:42:56.482911    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.351673    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 18:02:17.427619831 +0000 UTC
Jan 22 09:42:57 crc kubenswrapper[4824]: E0122 09:42:57.476587    4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.483088    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.484389    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.484462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.484488    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.957654    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.957853    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.958963    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.959027    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:57 crc kubenswrapper[4824]: I0122 09:42:57.959039    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.213022    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.352795    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 09:10:37.231346781 +0000 UTC
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.486001    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.487208    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.487236    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.487247    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:42:58 crc kubenswrapper[4824]: I0122 09:42:58.490383    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:59 crc kubenswrapper[4824]: I0122 09:42:59.134590    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:42:59 crc kubenswrapper[4824]: I0122 09:42:59.352948    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 06:11:57.821926309 +0000 UTC
Jan 22 09:42:59 crc kubenswrapper[4824]: I0122 09:42:59.489031    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:42:59 crc kubenswrapper[4824]: I0122 09:42:59.491613    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:42:59 crc kubenswrapper[4824]: I0122 09:42:59.491669    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:42:59 crc kubenswrapper[4824]: I0122 09:42:59.491702    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:00 crc kubenswrapper[4824]: W0122 09:43:00.103327    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.103740    4824 trace.go:236] Trace[704160526]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:50.102) (total time: 10001ms):
Jan 22 09:43:00 crc kubenswrapper[4824]: Trace[704160526]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:43:00.103)
Jan 22 09:43:00 crc kubenswrapper[4824]: Trace[704160526]: [10.001538652s] [10.001538652s] END
Jan 22 09:43:00 crc kubenswrapper[4824]: E0122 09:43:00.103770    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError"
Jan 22 09:43:00 crc kubenswrapper[4824]: W0122 09:43:00.133296    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.133466    4824 trace.go:236] Trace[847789836]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:50.131) (total time: 10001ms):
Jan 22 09:43:00 crc kubenswrapper[4824]: Trace[847789836]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:43:00.133)
Jan 22 09:43:00 crc kubenswrapper[4824]: Trace[847789836]: [10.001571305s] [10.001571305s] END
Jan 22 09:43:00 crc kubenswrapper[4824]: E0122 09:43:00.133509    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError"
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.335519    4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.353120    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 16:01:10.392319248 +0000 UTC
Jan 22 09:43:00 crc kubenswrapper[4824]: E0122 09:43:00.358471    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s"
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.491724    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.492652    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.492694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:00 crc kubenswrapper[4824]: I0122 09:43:00.492703    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:00 crc kubenswrapper[4824]: E0122 09:43:00.587836    4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc"
Jan 22 09:43:01 crc kubenswrapper[4824]: W0122 09:43:01.022089    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.022228    4824 trace.go:236] Trace[336049679]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:51.021) (total time: 10001ms):
Jan 22 09:43:01 crc kubenswrapper[4824]: Trace[336049679]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (09:43:01.022)
Jan 22 09:43:01 crc kubenswrapper[4824]: Trace[336049679]: [10.00107449s] [10.00107449s] END
Jan 22 09:43:01 crc kubenswrapper[4824]: E0122 09:43:01.022263    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError"
Jan 22 09:43:01 crc kubenswrapper[4824]: W0122 09:43:01.165323    4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.165557    4824 trace.go:236] Trace[1576130429]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Jan-2026 09:42:51.163) (total time: 10001ms):
Jan 22 09:43:01 crc kubenswrapper[4824]: Trace[1576130429]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:43:01.165)
Jan 22 09:43:01 crc kubenswrapper[4824]: Trace[1576130429]: [10.001926346s] [10.001926346s] END
Jan 22 09:43:01 crc kubenswrapper[4824]: E0122 09:43:01.165607    4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError"
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.353710    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 16:41:22.601001926 +0000 UTC
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.876480    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc"
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.876836    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.879002    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.879110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.879134    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:01 crc kubenswrapper[4824]: I0122 09:43:01.921350    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc"
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.135147    4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body=
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.135292    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)"
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.353903    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 18:45:02.313668296 +0000 UTC
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.496538    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.497387    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.497422    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.497436    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:02 crc kubenswrapper[4824]: I0122 09:43:02.513705    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.347480    4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body=
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.347750    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.354619    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 23:28:01.283520071 +0000 UTC
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.499604    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.501023    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.501116    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.501143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.788674    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.790290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.790335    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.790346    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:03 crc kubenswrapper[4824]: I0122 09:43:03.790401    4824 kubelet_node_status.go:76] "Attempting to register node" node="crc"
Jan 22 09:43:03 crc kubenswrapper[4824]: E0122 09:43:03.836014    4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError"
Jan 22 09:43:04 crc kubenswrapper[4824]: I0122 09:43:04.355769    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 07:28:29.508418557 +0000 UTC
Jan 22 09:43:05 crc kubenswrapper[4824]: I0122 09:43:05.356839    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:21:28.873978998 +0000 UTC
Jan 22 09:43:06 crc kubenswrapper[4824]: I0122 09:43:06.357399    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 20:14:11.764755577 +0000 UTC
Jan 22 09:43:06 crc kubenswrapper[4824]: I0122 09:43:06.574972    4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403}
Jan 22 09:43:06 crc kubenswrapper[4824]: I0122 09:43:06.575055    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403"
Jan 22 09:43:07 crc kubenswrapper[4824]: I0122 09:43:07.358378    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 15:53:05.474307868 +0000 UTC
Jan 22 09:43:07 crc kubenswrapper[4824]: E0122 09:43:07.476756    4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.353854    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.354047    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.355414    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.355462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.355473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.359508    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 12:23:43.562242719 +0000 UTC
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.360083    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.512780    4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.513063    4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.514220    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.514274    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:08 crc kubenswrapper[4824]: I0122 09:43:08.514286    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:09 crc kubenswrapper[4824]: I0122 09:43:09.360164    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 16:01:25.214887088 +0000 UTC
Jan 22 09:43:10 crc kubenswrapper[4824]: I0122 09:43:10.361080    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 13:27:34.045952638 +0000 UTC
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.361702    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 15:16:12.046559741 +0000 UTC
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.555186    4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.575584    4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.576334    4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.579496    4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.590707    4824 kubelet_node_status.go:115] "Node was previously registered" node="crc"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.591281    4824 kubelet_node_status.go:79] "Successfully registered node" node="crc"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.592925    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.592953    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.592963    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.592977    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.592987    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.596585    4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43930->192.168.126.11:17697: read: connection reset by peer" start-of-body=
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.596585    4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43946->192.168.126.11:17697: read: connection reset by peer" start-of-body=
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.597474    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43946->192.168.126.11:17697: read: connection reset by peer"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.596830    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43930->192.168.126.11:17697: read: connection reset by peer"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.598100    4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body=
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.599271    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused"
Jan 22 09:43:11 crc kubenswrapper[4824]: E0122 09:43:11.618602    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.627052    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.627349    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.627462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.627571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.627663    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: E0122 09:43:11.640990    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.644211    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.648270    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.648329    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.648341    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.648382    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.648397    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.651726    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:43:11 crc kubenswrapper[4824]: E0122 09:43:11.659235    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.662746    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.662801    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.662817    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.662849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.662865    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: E0122 09:43:11.671920    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.675346    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.675415    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.675431    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.675508    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.675546    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: E0122 09:43:11.685809    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:11 crc kubenswrapper[4824]: E0122 09:43:11.686014    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.687972    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.688011    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.688036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.688066    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.688080    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.793292    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.793387    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.793402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.793423    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.793524    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.896058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.896441    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.896538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.896636    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.896731    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.968354    4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.999419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.999459    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.999468    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.999481    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:11 crc kubenswrapper[4824]: I0122 09:43:11.999492    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:11Z","lastTransitionTime":"2026-01-22T09:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.075009    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.101787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.102037    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.102116    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.102186    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.102244    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.102286    4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.140403    4824 csr.go:261] certificate signing request csr-5kgn8 is approved, waiting to be issued
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.160096    4824 csr.go:257] certificate signing request csr-5kgn8 is issued
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.204799    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.204840    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.204849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.204866    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.204876    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.307743    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.307793    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.307806    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.307828    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.307842    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.344075    4824 apiserver.go:52] "Watching apiserver"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.346297    4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.346560    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.346948    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.347098    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.347199    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.347490    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.347585    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.347801    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.347918    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.347972    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.348080    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.349991    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.350443    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.350623    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.350699    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.351021    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.351386    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.352321    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.357707    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.359864    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.362160    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 07:03:08.177179114 +0000 UTC
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.379818    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.388822    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.399855    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.408415    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.410092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.410202    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.410299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.410382    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.410447    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.421103    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.429058    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.438013    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.446545    4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.450203    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.462108    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484321    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484399    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484420    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484439    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484475    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484496    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484509    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484526    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484558    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484572    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484587    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484605    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484635    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484650    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484664    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484679    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484710    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484735    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484749    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484785    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484804    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484820    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484833    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484863    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484884    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484898    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484911    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484940    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484955    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484969    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.484984    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485013    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485028    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485042    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485058    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485074    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485104    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485121    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485143    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485174    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485193    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485206    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485232    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485264    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485277    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485292    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485305    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485335    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485349    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485392    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485406    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485423    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485455    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485470    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485484    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485627    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485647    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485661    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485676    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485718    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485733    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485747    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485781    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485796    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485809    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485824    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485854    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485869    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485885    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485899    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485929    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485945    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485958    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.485973    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486004    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486021    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486036    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486050    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486082    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486097    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486111    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486126    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486142    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486157    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486187    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486202    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486218    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486235    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486249    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486268    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486289    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486324    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486377    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486402    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486417    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486432    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486447    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486464    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486496    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486513    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486548    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486563    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486579    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486593    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486608    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486622    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486637    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486653    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486668    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486684    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486701    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486717    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486737    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486759    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486788    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486812    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486834    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486857    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.486880    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487304    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487300    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487307    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487523    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487533    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487523    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487535    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487697    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487819    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487897    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.487977    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488034    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488155    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488190    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488201    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488226    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488285    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488394    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488463    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488492    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488510    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488575    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488655    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488876    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488924    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.488964    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489046    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489069    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489085    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489120    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489169    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489172    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489225    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489293    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489404    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489418    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489533    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489657    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489699    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489741    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489752    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489928    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489969    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490005    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490014    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490068    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490186    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490209    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490460    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490646    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490844    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.490869    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.491103    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.489840    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.491615    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.491698    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.491706    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.491797    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.492206    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493127    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493289    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493504    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493583    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493739    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493793    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493810    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.493790    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.494162    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:12.994131106 +0000 UTC m=+25.897608099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.494233    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.494457    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.494511    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.494659    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.495036    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.495127    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.495334    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.495567    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.495704    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.496137    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.496311    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.497164    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.497202    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.498719    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.498725    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499114    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499190    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499517    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499570    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499651    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499691    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.499705    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500009    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500453    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500471    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500411    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500541    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500545    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500806    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500860    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.500842    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501107    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501256    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501484    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501506    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501664    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.491709    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501710    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501750    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501778    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501800    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501822    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501842    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501862    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501881    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501908    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501930    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501950    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501971    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501992    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502014    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502050    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502073    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502096    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502120    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502143    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502167    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502193    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502215    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502233    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502249    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502267    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502284    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502313    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502333    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502349    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502432    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502452    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502468    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502483    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502498    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502513    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502528    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502542    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502557    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502572    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502588    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502605    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502620    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502635    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502652    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502668    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502686    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502703    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502721    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502738    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502759    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502781    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503533    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503560    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503584    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503665    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503703    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503729    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503753    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503778    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503954    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504349    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504396    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504421    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504445    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504469    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504494    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504519    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504543    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504569    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504595    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504622    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504647    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504672    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504696    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504722    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504751    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") "
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504828    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504862    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504891    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504920    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504945    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504969    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504995    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505044    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505071    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505098    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505121    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505150    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505173    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505201    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505277    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505296    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505306    4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505316    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505325    4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505335    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505344    4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505356    4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505413    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505429    4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505471    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505483    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505498    4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505510    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505521    4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505536    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505548    4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505561    4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505574    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505586    4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505607    4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505619    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505630    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505641    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505652    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505664    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505677    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505692    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505704    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505718    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505731    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505743    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505755    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505766    4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505777    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505789    4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505801    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505813    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505826    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505838    4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505850    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505863    4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505875    4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505887    4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505900    4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505913    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505925    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505938    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505952    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505965    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505977    4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505990    4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506001    4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506014    4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506027    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506039    4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506051    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506064    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506076    4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506088    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506100    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506113    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506126    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506139    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506151    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506163    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506176    4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506188    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506207    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506224    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506237    4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506249    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506262    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506274    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506286    4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506298    4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506310    4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506321    4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506334    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506371    4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506383    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506396    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506431    4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506443    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506457    4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506470    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506481    4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506493    4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506505    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506517    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506528    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506540    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506553    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506564    4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506577    4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506590    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506602    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506614    4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506627    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506638    4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506650    4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506663    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506674    4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506685    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.501880    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513953    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502109    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514092    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.502660    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.503961    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504213    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504567    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504685    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.504743    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505317    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505345    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505506    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.505831    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506505    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.506990    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.507100    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.507482    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.507774    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.507787    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.507801    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.508235    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.508256    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.508721    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.508740    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.508738    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.508777    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.509269    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.509307    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.509648    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.509841    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510040    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510093    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510096    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510263    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510397    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510559    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.510764    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.511286    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.511532    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.511799    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.511911    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.511929    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512183    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512205    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512223    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512261    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512425    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512520    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.512654    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513029    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.513426    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.514539    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:13.014510973 +0000 UTC m=+25.917987966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514684    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514698    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513614    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513744    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513867    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513884    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.513986    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514038    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514158    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.514160    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514238    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514386    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514814    4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.514916    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515065    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515178    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515176    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.515184    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:13.015160772 +0000 UTC m=+25.918637775 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515268    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515410    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515468    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515541    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515635    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515678    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515695    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515703    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515718    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.515727    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.516624    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.516832    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.517087    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.517850    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.519377    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.520294    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.520649    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.523196    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.526892    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.528255    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.528888    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.528917    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.528930    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.529007    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:13.028990404 +0000 UTC m=+25.932467397 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.530251    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.530315    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.530790    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.530930    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.531430    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.534251    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.537244    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.537419    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.537466    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.540550    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.541026    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.545315    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mdbj5"]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.545738    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.546106    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.546142    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.546158    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.546234    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:13.046205817 +0000 UTC m=+25.949682810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.547536    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.548930    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.549231    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.549398    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.550052    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.552231    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.553937    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.558921    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.560975    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.564481    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe" exitCode=255
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.565568    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.566610    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.580154    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.594865    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: E0122 09:43:12.594893    4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608318    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608594    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/798e70c0-4438-46c9-9efc-be9121d1e60e-hosts-file\") pod \"node-resolver-mdbj5\" (UID: \"798e70c0-4438-46c9-9efc-be9121d1e60e\") " pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608644    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwdch\" (UniqueName: \"kubernetes.io/projected/798e70c0-4438-46c9-9efc-be9121d1e60e-kube-api-access-zwdch\") pod \"node-resolver-mdbj5\" (UID: \"798e70c0-4438-46c9-9efc-be9121d1e60e\") " pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608690    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608812    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608832    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608854    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608873    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608886    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608898    4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608915    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608927    4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608945    4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608957    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608975    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608987    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.608999    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609015    4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609027    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609039    4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609050    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609065    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609077    4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609088    4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609100    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609116    4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609128    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609141    4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609152    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609168    4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609181    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609204    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609222    4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609233    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609245    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609263    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609278    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609290    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609305    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609317    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609332    4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609343    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609376    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609392    4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609403    4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609414    4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609425    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609441    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609460    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609471    4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609483    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609497    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609512    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609532    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609547    4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609562    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609573    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609584    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609599    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609618    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609629    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609641    4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609656    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609670    4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609682    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609694    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609709    4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609720    4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609742    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609754    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609768    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609779    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609790    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609804    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609819    4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609831    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609845    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609859    4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609871    4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609882    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.609893    4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610096    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610113    4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610127    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610146    4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610157    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610168    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610179    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610194    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610204    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610215    4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610228    4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610242    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610253    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.610461    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.615531    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.621398    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.622492    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.623175    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.626535    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.626587    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.626601    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.629833    4824 scope.go:117] "RemoveContainer" containerID="86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.630555    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.636095    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.646049    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.664653    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.665439    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.671705    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.679815    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.680516    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.697816    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.714894    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/798e70c0-4438-46c9-9efc-be9121d1e60e-hosts-file\") pod \"node-resolver-mdbj5\" (UID: \"798e70c0-4438-46c9-9efc-be9121d1e60e\") " pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.715045    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwdch\" (UniqueName: \"kubernetes.io/projected/798e70c0-4438-46c9-9efc-be9121d1e60e-kube-api-access-zwdch\") pod \"node-resolver-mdbj5\" (UID: \"798e70c0-4438-46c9-9efc-be9121d1e60e\") " pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.715096    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/798e70c0-4438-46c9-9efc-be9121d1e60e-hosts-file\") pod \"node-resolver-mdbj5\" (UID: \"798e70c0-4438-46c9-9efc-be9121d1e60e\") " pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.718588    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.731667    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.735025    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.735054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.735065    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.735082    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.735094    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.738309    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwdch\" (UniqueName: \"kubernetes.io/projected/798e70c0-4438-46c9-9efc-be9121d1e60e-kube-api-access-zwdch\") pod \"node-resolver-mdbj5\" (UID: \"798e70c0-4438-46c9-9efc-be9121d1e60e\") " pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.744250    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.764257    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.778664    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.799831    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.808905    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.822346    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.843040    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.843090    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.843104    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.843127    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.843139    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.844340    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.902295    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mdbj5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.939503    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-zvvt5"]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.939773    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9wczc"]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.940173    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4527v"]
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.940423    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.940738    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.941427    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:12 crc kubenswrapper[4824]: W0122 09:43:12.943834    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod798e70c0_4438_46c9_9efc_be9121d1e60e.slice/crio-c555808473e33636a238502311f8f94b392cf35cdf3d7a31bad6201cf016ef8b WatchSource:0}: Error finding container c555808473e33636a238502311f8f94b392cf35cdf3d7a31bad6201cf016ef8b: Status 404 returned error can't find the container with id c555808473e33636a238502311f8f94b392cf35cdf3d7a31bad6201cf016ef8b
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.945408    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.945446    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.945459    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.945481    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.945492    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:12Z","lastTransitionTime":"2026-01-22T09:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947165    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947169    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947194    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947232    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947165    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947487    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947720    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947765    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.947574    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.948108    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.948145    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.953456    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.973039    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:12 crc kubenswrapper[4824]: I0122 09:43:12.988525    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.005099    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.018916    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019075    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-cnibin\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019112    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019137    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-system-cni-dir\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019166    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t72pd\" (UniqueName: \"kubernetes.io/projected/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-kube-api-access-t72pd\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019189    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-system-cni-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019219    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-os-release\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019242    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-multus-certs\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019277    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019299    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019344    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22c2c99-5647-4fc8-9550-7e2a82584d55-mcd-auth-proxy-config\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019386    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-cni-binary-copy\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019408    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-cni-bin\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019444    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-daemon-config\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019465    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqtpk\" (UniqueName: \"kubernetes.io/projected/e22c2c99-5647-4fc8-9550-7e2a82584d55-kube-api-access-bqtpk\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019490    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-cni-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019514    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-hostroot\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019537    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e22c2c99-5647-4fc8-9550-7e2a82584d55-proxy-tls\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019560    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-os-release\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019581    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-cni-multus\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019603    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/68893d7c-efac-44f9-8935-843169e413ad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019622    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-etc-kubernetes\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019641    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-cnibin\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019661    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e22c2c99-5647-4fc8-9550-7e2a82584d55-rootfs\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019680    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-netns\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019698    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-kubelet\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019718    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-conf-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019739    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj5rl\" (UniqueName: \"kubernetes.io/projected/68893d7c-efac-44f9-8935-843169e413ad-kube-api-access-qj5rl\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019761    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-socket-dir-parent\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019781    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-k8s-cni-cncf-io\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.019803    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/68893d7c-efac-44f9-8935-843169e413ad-cni-binary-copy\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.019922    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:14.019903045 +0000 UTC m=+26.923380038 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.020004    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.020048    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:14.020038719 +0000 UTC m=+26.923515712 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.020722    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.020843    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:14.020831172 +0000 UTC m=+26.924308375 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.025224    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.038177    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.047235    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.052257    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.052561    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.052897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.053284    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.053506    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.060912    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.071288    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.080734    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.095111    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.106739    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.119018    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120193    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-system-cni-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120244    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-os-release\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120407    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-system-cni-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120434    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-os-release\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120318    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-multus-certs\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120499    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t72pd\" (UniqueName: \"kubernetes.io/projected/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-kube-api-access-t72pd\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120542    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120564    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120566    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-multus-certs\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120591    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22c2c99-5647-4fc8-9550-7e2a82584d55-mcd-auth-proxy-config\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120615    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-cni-binary-copy\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120638    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-cni-bin\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120657    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-cni-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120674    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-daemon-config\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120697    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqtpk\" (UniqueName: \"kubernetes.io/projected/e22c2c99-5647-4fc8-9550-7e2a82584d55-kube-api-access-bqtpk\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120714    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-os-release\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120870    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.120867    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-cni-bin\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.121012    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.121029    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.121043    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.121097    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:14.121075008 +0000 UTC m=+27.024551991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121532    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-cni-multus\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121605    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-hostroot\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121641    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e22c2c99-5647-4fc8-9550-7e2a82584d55-proxy-tls\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121648    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-cni-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121815    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e22c2c99-5647-4fc8-9550-7e2a82584d55-mcd-auth-proxy-config\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121829    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-daemon-config\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121899    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-os-release\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121899    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-cni-multus\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121929    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-hostroot\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121671    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/68893d7c-efac-44f9-8935-843169e413ad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.121985    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-cni-binary-copy\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122022    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-etc-kubernetes\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122052    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e22c2c99-5647-4fc8-9550-7e2a82584d55-rootfs\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122075    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-cnibin\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122123    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e22c2c99-5647-4fc8-9550-7e2a82584d55-rootfs\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122130    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-etc-kubernetes\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122101    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj5rl\" (UniqueName: \"kubernetes.io/projected/68893d7c-efac-44f9-8935-843169e413ad-kube-api-access-qj5rl\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122195    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-socket-dir-parent\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122221    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-k8s-cni-cncf-io\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122245    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-netns\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122267    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-socket-dir-parent\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122271    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-kubelet\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122166    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-cnibin\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122295    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-conf-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122305    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-netns\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122328    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-run-k8s-cni-cncf-io\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122343    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/68893d7c-efac-44f9-8935-843169e413ad-cni-binary-copy\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122378    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-multus-conf-dir\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122362    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-host-var-lib-kubelet\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122411    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.122505    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122514    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/68893d7c-efac-44f9-8935-843169e413ad-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.122529    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.122545    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122544    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-system-cni-dir\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122580    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-cnibin\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: E0122 09:43:13.122596    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:14.122578222 +0000 UTC m=+27.026055385 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122651    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-system-cni-dir\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122671    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/68893d7c-efac-44f9-8935-843169e413ad-cnibin\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.122809    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/68893d7c-efac-44f9-8935-843169e413ad-cni-binary-copy\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.127890    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e22c2c99-5647-4fc8-9550-7e2a82584d55-proxy-tls\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.138142    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.142736    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t72pd\" (UniqueName: \"kubernetes.io/projected/00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c-kube-api-access-t72pd\") pod \"multus-zvvt5\" (UID: \"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\") " pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.146420    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj5rl\" (UniqueName: \"kubernetes.io/projected/68893d7c-efac-44f9-8935-843169e413ad-kube-api-access-qj5rl\") pod \"multus-additional-cni-plugins-9wczc\" (UID: \"68893d7c-efac-44f9-8935-843169e413ad\") " pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.147071    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqtpk\" (UniqueName: \"kubernetes.io/projected/e22c2c99-5647-4fc8-9550-7e2a82584d55-kube-api-access-bqtpk\") pod \"machine-config-daemon-4527v\" (UID: \"e22c2c99-5647-4fc8-9550-7e2a82584d55\") " pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.157037    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.157106    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.157117    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.157154    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.157170    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.159008    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.160922    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-22 09:38:12 +0000 UTC, rotation deadline is 2026-10-06 12:24:44.256525265 +0000 UTC
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.161000    4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6170h41m31.095526888s for next certificate rotation
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.173192    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.182986    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.194278    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.208923    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.224038    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.237978    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.251154    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.260479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.260533    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.260550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.260570    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.260584    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.266690    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:43:13 crc kubenswrapper[4824]: W0122 09:43:13.278500    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode22c2c99_5647_4fc8_9550_7e2a82584d55.slice/crio-e7f62be8211b4fb5a6510cbc2600419b6a74692a955f6a508f8b925c73905664 WatchSource:0}: Error finding container e7f62be8211b4fb5a6510cbc2600419b6a74692a955f6a508f8b925c73905664: Status 404 returned error can't find the container with id e7f62be8211b4fb5a6510cbc2600419b6a74692a955f6a508f8b925c73905664
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.285797    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.292521    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-zvvt5"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.303492    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9wczc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.309297    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8vt4"]
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.310430    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.312979    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.313631    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.313739    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.313826    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.313940    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.313974    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.314647    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.315296    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.328644    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.346944    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.361740    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.362465    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 05:12:13.771361087 +0000 UTC
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.362754    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.362809    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.362818    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.362850    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.362861    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.382881    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.400699    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.409805    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.410424    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.411892    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.412578    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.413694    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.414287    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.414862    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.416654    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.417415    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.420416    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.420955    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.421741    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.422797    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.423309    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.424219    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.424864    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425536    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425579    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-kubelet\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425598    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-var-lib-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425618    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-bin\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425657    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-ovn\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425677    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-netd\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425693    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-env-overrides\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425711    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovn-node-metrics-cert\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425738    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425767    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-slash\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425789    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-log-socket\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.425962    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-script-lib\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426026    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pq4s\" (UniqueName: \"kubernetes.io/projected/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-kube-api-access-5pq4s\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426059    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-config\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426086    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-systemd-units\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426102    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-netns\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426116    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-ovn-kubernetes\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426151    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-systemd\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426185    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-node-log\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426204    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-etc-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.426636    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.427709    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.428860    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.429441    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.430080    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.431158    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.431743    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.432761    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.436018    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.439696    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.440314    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.449790    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.450333    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.451598    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.452200    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.452812    4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.452941    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.455167    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.455807    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.456866    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.461926    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.462863    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.466535    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.467392    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.467595    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.467635    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.467645    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.467664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.467678    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.468075    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.468935    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.473049    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.474150    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.474791    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.475714    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.476242    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.477118    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.477874    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.478840    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.479314    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.479887    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.483193    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.484004    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.484940    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.491665    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.524135    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.526583    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-kubelet\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.526703    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-var-lib-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.526801    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-bin\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.526923    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-ovn\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527012    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-netd\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527095    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-env-overrides\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527173    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovn-node-metrics-cert\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527276    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527400    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-slash\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527485    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-log-socket\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527553    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-script-lib\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527596    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527543    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-var-lib-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527576    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-bin\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527680    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-netd\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527754    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pq4s\" (UniqueName: \"kubernetes.io/projected/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-kube-api-access-5pq4s\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527819    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-config\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527884    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-systemd-units\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527952    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-netns\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528016    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-ovn-kubernetes\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528089    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-systemd\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528153    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-node-log\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528222    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-etc-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528299    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528444    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528531    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-slash\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528606    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-log-socket\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527495    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-ovn\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528813    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-env-overrides\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.528959    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-netns\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.527522    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-kubelet\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529148    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-ovn-kubernetes\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529227    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-systemd\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529252    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-node-log\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529280    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-etc-openvswitch\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529308    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-systemd-units\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529605    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-script-lib\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.529833    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-config\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.536853    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovn-node-metrics-cert\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.541633    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.562812    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pq4s\" (UniqueName: \"kubernetes.io/projected/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-kube-api-access-5pq4s\") pod \"ovnkube-node-s8vt4\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") " pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.576277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.576318    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.576331    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.576348    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.576379    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.581034    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerStarted","Data":"660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.581070    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerStarted","Data":"577965336ac445dcd5c53c96bea10ec55988b7fe8bd9fc74300d19b5f71b62b9"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.583745    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.583799    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.583809    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"e7f62be8211b4fb5a6510cbc2600419b6a74692a955f6a508f8b925c73905664"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.587020    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.587082    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"693a7a2ae49289773d6150545ba747337cda0092ffc6ff500af11ba30cab515a"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.593000    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.595247    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.595470    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.595700    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.600940    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerStarted","Data":"b7bff41750c266e495c958a5b316514af20e5c6bab43c104f09f1a716ff89ce7"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.609180    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.609210    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.609222    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d504538cb201811e2b879f4da373edf9b8cc3e0a934fd2df417c2ed63453f951"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.617077    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mdbj5" event={"ID":"798e70c0-4438-46c9-9efc-be9121d1e60e","Type":"ContainerStarted","Data":"49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.617102    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mdbj5" event={"ID":"798e70c0-4438-46c9-9efc-be9121d1e60e","Type":"ContainerStarted","Data":"c555808473e33636a238502311f8f94b392cf35cdf3d7a31bad6201cf016ef8b"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.618564    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"810d7dfda34cb569122d9f87e8797758464da4de2065525fb0c9fbb4e3e7340d"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.623457    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.648632    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.679814    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.679868    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.679880    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.679900    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.679912    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.683201    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.717782    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.757113    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.782289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.782339    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.782380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.782402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.782413    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.795497    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.861529    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.876979    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.884648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.884688    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.884697    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.884716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.884726    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.916621    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.959669    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.988486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.988856    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.988868    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.988885    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:13 crc kubenswrapper[4824]: I0122 09:43:13.988894    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:13Z","lastTransitionTime":"2026-01-22T09:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.005697    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.036289    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.049285    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.049424    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.049512    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.049579    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:16.049540839 +0000 UTC m=+28.953017832 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.049627    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.049650    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.049737    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:16.049710144 +0000 UTC m=+28.953187207 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.049820    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:16.049802057 +0000 UTC m=+28.953279050 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.091550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.091591    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.091600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.091616    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.091626    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.106036    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.150449    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.150509    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150646    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150664    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150678    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150733    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:16.150717633 +0000 UTC m=+29.054194616 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150756    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150799    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150816    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.150896    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:16.150872797 +0000 UTC m=+29.054349820 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.152256    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.177222    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.193709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.193758    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.193770    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.193787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.193798    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.206368    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.231574    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.273473    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.296560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.296622    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.296639    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.296664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.296680    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.363447    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 17:28:53.67257963 +0000 UTC
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.398914    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.398955    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.398967    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.398993    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.399006    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.403339    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.403386    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.403456    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.403339    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.403537    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:14 crc kubenswrapper[4824]: E0122 09:43:14.403580    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.501955    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.502001    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.502012    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.502028    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.502039    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.604886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.605200    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.605210    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.605225    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.605235    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.622673    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35" exitCode=0
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.622739    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.622777    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"24ee1eb98143886fb2956ccf86fee1e5aab73aa59d6d3fbd9bba7ab3b071fc3d"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.624551    4824 generic.go:334] "Generic (PLEG): container finished" podID="68893d7c-efac-44f9-8935-843169e413ad" containerID="9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55" exitCode=0
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.624584    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerDied","Data":"9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.639877    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.665244    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.679459    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.692474    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.704708    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.707346    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.707409    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.707419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.707435    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.707444    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.718875    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.736493    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.755445    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.778955    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.793005    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.804222    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.813408    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.813455    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.813464    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.813481    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.813493    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.825291    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.840384    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.855739    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.874647    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.916392    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.916432    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.916465    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.916486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.916498    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:14Z","lastTransitionTime":"2026-01-22T09:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.916804    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.961302    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:14 crc kubenswrapper[4824]: I0122 09:43:14.997606    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:14Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.019119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.019491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.019503    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.019522    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.019534    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.041403    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.074871    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.114465    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.122215    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.122261    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.122270    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.122288    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.122299    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.155561    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.194954    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.224608    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.224672    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.224683    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.224702    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.224720    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.236178    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.273522    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.315579    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.327716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.327757    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.327765    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.327783    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.327793    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.363697    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 11:06:11.915785864 +0000 UTC
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.430741    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.430795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.430810    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.430829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.430851    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.533427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.533474    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.533487    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.533504    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.533513    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.604928    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tqrnv"]
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.605560    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.608662    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.608780    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.608809    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.609291    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.620922    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.636915    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.636966    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.636977    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.636995    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.637008    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.637928    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.640007    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.640062    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.640073    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.640083    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.640093    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.640101    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.642010    4824 generic.go:334] "Generic (PLEG): container finished" podID="68893d7c-efac-44f9-8935-843169e413ad" containerID="1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd" exitCode=0
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.642082    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerDied","Data":"1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.645005    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.652992    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.672291    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.687235    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.700167    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.714708    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.735737    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.739381    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.739429    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.739439    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.739456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.739466    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.755648    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.767659    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cced567c-003b-445f-b568-045b5a667853-serviceca\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.767770    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cced567c-003b-445f-b568-045b5a667853-host\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.767826    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-896ns\" (UniqueName: \"kubernetes.io/projected/cced567c-003b-445f-b568-045b5a667853-kube-api-access-896ns\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.797658    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.834536    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.842139    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.842177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.842193    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.842214    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.842225    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.868527    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cced567c-003b-445f-b568-045b5a667853-serviceca\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.868612    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cced567c-003b-445f-b568-045b5a667853-host\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.868644    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-896ns\" (UniqueName: \"kubernetes.io/projected/cced567c-003b-445f-b568-045b5a667853-kube-api-access-896ns\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.868838    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cced567c-003b-445f-b568-045b5a667853-host\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.873134    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/cced567c-003b-445f-b568-045b5a667853-serviceca\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.878638    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.904615    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-896ns\" (UniqueName: \"kubernetes.io/projected/cced567c-003b-445f-b568-045b5a667853-kube-api-access-896ns\") pod \"node-ca-tqrnv\" (UID: \"cced567c-003b-445f-b568-045b5a667853\") " pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.923534    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tqrnv"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.944243    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.945917    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.945949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.945957    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.945972    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.945983    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:15Z","lastTransitionTime":"2026-01-22T09:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:15 crc kubenswrapper[4824]: I0122 09:43:15.978488    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:15Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.030622    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.049298    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.049374    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.049389    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.049409    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.049421    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.055677    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.071903    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.072026    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.072097    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:20.072063134 +0000 UTC m=+32.975540157 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.072103    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.072147    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.072167    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:20.072157526 +0000 UTC m=+32.975634629 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.072245    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.072311    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:20.07229052 +0000 UTC m=+32.975767603 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.095431    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.139672    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.151831    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.151897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.151912    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.151957    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.151970    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.174071    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.174310    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174468    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174465    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174509    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174522    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174590    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:20.174568836 +0000 UTC m=+33.078045829 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174487    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174690    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.174892    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:20.174787183 +0000 UTC m=+33.078264176 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.176500    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.213245    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.254012    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.254710    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.254751    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.254765    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.254783    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.254795    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.293486    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.333332    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.359690    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.359747    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.359760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.359782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.359803    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.363925    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 12:58:35.599852932 +0000 UTC
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.375311    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.403405    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.403559    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.403720    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.403910    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.403348    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:16 crc kubenswrapper[4824]: E0122 09:43:16.404043    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.420142    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.456294    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.462213    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.462515    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.462578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.462668    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.462771    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.495801    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.536788    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.565402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.565719    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.565806    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.565900    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.565997    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.651200    4824 generic.go:334] "Generic (PLEG): container finished" podID="68893d7c-efac-44f9-8935-843169e413ad" containerID="877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa" exitCode=0
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.651276    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerDied","Data":"877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.652705    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tqrnv" event={"ID":"cced567c-003b-445f-b568-045b5a667853","Type":"ContainerStarted","Data":"9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.652750    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tqrnv" event={"ID":"cced567c-003b-445f-b568-045b5a667853","Type":"ContainerStarted","Data":"9fac83e805c2c31e596e402d1e9dac254d29225aaf0fc7c19c60d77fde7e8895"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.669168    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.669430    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.669582    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.669717    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.670287    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.670583    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.688060    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.704688    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.725589    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.738973    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.774585    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.774887    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.774949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.774965    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.774987    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.775004    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.815332    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.852482    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.877708    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.877753    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.877763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.877780    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.877790    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.894393    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.932491    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.975483    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.980102    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.980190    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.980207    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.980229    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:16 crc kubenswrapper[4824]: I0122 09:43:16.980255    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:16Z","lastTransitionTime":"2026-01-22T09:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.015950    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.056166    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.082517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.082554    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.082562    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.082577    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.082586    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.097958    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.135334    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.176858    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.184559    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.184602    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.184612    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.184629    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.184641    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.215141    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.254058    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.257056    4824 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.286586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.286632    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.286641    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.286658    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.286670    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.296242    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.338549    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.364397    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 14:33:50.138540503 +0000 UTC
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.373735    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.388786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.388842    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.388853    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.388873    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.388885    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.414955    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.454930    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.490709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.490993    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.491076    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.491152    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.491224    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.493560    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.533287    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.574188    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.593105    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.593518    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.593641    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.593760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.593918    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.621676    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.654397    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.658741    4824 generic.go:334] "Generic (PLEG): container finished" podID="68893d7c-efac-44f9-8935-843169e413ad" containerID="5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c" exitCode=0
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.658788    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerDied","Data":"5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.696365    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.696410    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.696424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.696442    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.696456    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.704519    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.750057    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.776959    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.800838    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.800886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.800899    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.800919    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.800938    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.813914    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.859484    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.895251    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.903100    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.903160    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.903169    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.903187    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.903217    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:17Z","lastTransitionTime":"2026-01-22T09:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.936027    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:17 crc kubenswrapper[4824]: I0122 09:43:17.979465    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.006018    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.006063    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.006072    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.006088    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.006098    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.019537    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.058562    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.095848    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.109393    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.109447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.109460    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.109479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.109493    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.134665    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.173468    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.211885    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.211936    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.211945    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.211961    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.211972    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.215670    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.255808    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.295051    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.316614    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.316679    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.316704    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.316727    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.316740    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.333552    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.365521    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:42:02.632805279 +0000 UTC
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.374603    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.403315    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.403402    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.403401    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:18 crc kubenswrapper[4824]: E0122 09:43:18.403474    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:18 crc kubenswrapper[4824]: E0122 09:43:18.403555    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:18 crc kubenswrapper[4824]: E0122 09:43:18.403610    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.416220    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.418903    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.418960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.418970    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.418991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.419004    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.454077    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.495278    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.522494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.522879    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.522897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.522917    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.522929    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.539031    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.574624    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.612002    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.625068    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.625113    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.625124    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.625151    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.625163    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.654079    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.663964    4824 generic.go:334] "Generic (PLEG): container finished" podID="68893d7c-efac-44f9-8935-843169e413ad" containerID="abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87" exitCode=0
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.664075    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerDied","Data":"abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.667691    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.697281    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.731058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.731099    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.731111    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.731129    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.731140    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.734530    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.773981    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.817887    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.833418    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.833457    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.833467    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.833483    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:18 crc kubenswrapper[4824]: I0122 09:43:18.833493    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.856688    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.895791    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.935708    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.935734    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.935754    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.935770    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.935778    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:18Z","lastTransitionTime":"2026-01-22T09:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.942064    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:18.974245    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.012876    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.038339    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.038396    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.038408    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.038423    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.038432    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.056684    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.099259    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.134079    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.141034    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.141075    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.141084    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.141102    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.141113    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.172126    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.215131    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.243060    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.243102    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.243114    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.243132    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.243143    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.256625    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.294197    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.335788    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.345259    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.345300    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.345312    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.345329    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.345341    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.366590    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 21:46:29.064603911 +0000 UTC
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.448253    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.448349    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.448424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.448454    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.448476    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.551494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.551532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.551540    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.551555    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.551565    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.654398    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.654482    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.654517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.654548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.654566    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.674708    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerStarted","Data":"5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.687336    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.702443    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.719783    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.734000    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.747075    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.757653    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.757711    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.757724    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.757746    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.757763    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.765138    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.778700    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.792958    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.811561    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.823606    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.832106    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.843089    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.875742    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.875790    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.875799    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.875815    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.875825    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.881888    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.901414    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:19Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.978879    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.978936    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.978950    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.978968    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:19 crc kubenswrapper[4824]: I0122 09:43:19.978979    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:19Z","lastTransitionTime":"2026-01-22T09:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.082739    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.082806    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.082824    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.082849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.082867    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.112209    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.112329    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.112382    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.112490    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.112512    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.112528    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:28.112485825 +0000 UTC m=+41.015962858 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.112588    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:28.112565528 +0000 UTC m=+41.016042521 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.112607    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:28.112598809 +0000 UTC m=+41.016075802 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.186274    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.186333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.186345    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.186390    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.186404    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.213036    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.213099    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213276    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213303    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213317    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213277    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213427    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213443    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213394    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:28.21337358 +0000 UTC m=+41.116850583 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.213535    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:28.213493404 +0000 UTC m=+41.116970477 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.289251    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.289299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.289308    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.289325    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.289336    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.367531    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 00:27:29.047471419 +0000 UTC
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.391525    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.391575    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.391585    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.391600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.391612    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.403837    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.403935    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.403956    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.403984    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.404056    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:20 crc kubenswrapper[4824]: E0122 09:43:20.404130    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.500176    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.500237    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.500250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.500271    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.500285    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.602927    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.603006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.603016    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.603034    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.603045    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.682924    4824 generic.go:334] "Generic (PLEG): container finished" podID="68893d7c-efac-44f9-8935-843169e413ad" containerID="5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006" exitCode=0
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.682983    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerDied","Data":"5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.700761    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.705166    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.705231    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.705249    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.705277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.705290    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.721171    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.737333    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.753939    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.768747    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.788102    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.807679    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.810509    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.810555    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.810571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.810597    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.810612    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.826047    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.843372    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.859762    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.870990    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.884717    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.900055    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.912581    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.912606    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.912613    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.912628    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.912638    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:20Z","lastTransitionTime":"2026-01-22T09:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:20 crc kubenswrapper[4824]: I0122 09:43:20.914963    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:20Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.015047    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.015071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.015078    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.015092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.015100    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.118063    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.118140    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.118165    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.118200    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.118222    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.225913    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.225956    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.225964    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.225981    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.225992    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.327934    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.328006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.328022    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.328044    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.328059    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.368250    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 22:37:38.798854772 +0000 UTC
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.431402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.431455    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.431469    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.431488    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.431501    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.534256    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.534298    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.534310    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.534326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.534337    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.636845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.636883    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.636906    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.636924    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.636933    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.691611    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" event={"ID":"68893d7c-efac-44f9-8935-843169e413ad","Type":"ContainerStarted","Data":"c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.696102    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.696497    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.696531    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.696546    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.714307    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.724657    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.724735    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.737000    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.739171    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.739334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.739477    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.739583    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.739714    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.749112    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.761106    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.773321    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.784159    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.798289    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.809751    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.822565    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.835888    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.841645    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.841689    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.841701    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.841730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.841740    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.849143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.849197    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.849211    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.849231    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.849245    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.849322    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: E0122 09:43:21.863578    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.864763    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.867913    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.867953    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.867960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.867977    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.867986    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: E0122 09:43:21.881085    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.881696    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.885458    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.885483    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.885491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.885506    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.885515    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: E0122 09:43:21.898897    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.903709    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.905790    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.905818    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.905825    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.905841    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.905851    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: E0122 09:43:21.919812    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.921340    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.925549    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.925578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.925586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.925600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.925610    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.936493    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: E0122 09:43:21.939322    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: E0122 09:43:21.939533    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.944122    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.944162    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.944171    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.944187    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.944198    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:21Z","lastTransitionTime":"2026-01-22T09:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.947167    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.955613    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.965114    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.972212    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.980814    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:21 crc kubenswrapper[4824]: I0122 09:43:21.991242    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.003094    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:21Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.014286    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:22Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.023646    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:22Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.034141    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:22Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.046288    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.046334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.046343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.046375    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.046385    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.049314    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:22Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.065219    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:22Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.148653    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.148705    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.148716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.148736    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.148747    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.251438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.251512    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.251530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.251560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.251577    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.355133    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.355185    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.355196    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.355214    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.355228    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.368560    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 16:17:15.252368068 +0000 UTC
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.404018    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.404071    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:22 crc kubenswrapper[4824]: E0122 09:43:22.404235    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:22 crc kubenswrapper[4824]: E0122 09:43:22.404438    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.404523    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:22 crc kubenswrapper[4824]: E0122 09:43:22.404631    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.458281    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.458329    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.458339    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.458376    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.458387    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.561239    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.561411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.561431    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.561499    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.561522    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.664896    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.664948    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.664958    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.664981    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.664994    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.768093    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.768134    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.768143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.768162    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.768171    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.871335    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.871392    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.871400    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.871419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.871428    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.974336    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.974408    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.974418    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.974456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:22 crc kubenswrapper[4824]: I0122 09:43:22.974472    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:22Z","lastTransitionTime":"2026-01-22T09:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.078136    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.078176    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.078187    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.078205    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.078216    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.181851    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.181926    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.181939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.181960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.181974    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.284860    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.284904    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.284917    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.284937    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.284949    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.369220    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 23:43:26.910740836 +0000 UTC
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.387723    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.387786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.387803    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.387837    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.387855    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.490995    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.491070    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.491088    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.491114    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.491133    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.593881    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.593969    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.593998    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.594032    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.594057    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.697399    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.697472    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.697492    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.697517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.697536    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.801123    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.801196    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.801217    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.801244    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.801263    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.904487    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.904548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.904566    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.904588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:23 crc kubenswrapper[4824]: I0122 09:43:23.904605    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:23Z","lastTransitionTime":"2026-01-22T09:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.008118    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.008253    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.008281    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.008312    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.008332    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.111150    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.111189    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.111197    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.111213    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.111223    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.214412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.214488    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.214513    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.214544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.214563    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.317254    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.317328    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.317346    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.317401    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.317420    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.369840    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 21:01:49.774701848 +0000 UTC
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.403217    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.403349    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.403241    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:24 crc kubenswrapper[4824]: E0122 09:43:24.403595    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:24 crc kubenswrapper[4824]: E0122 09:43:24.403755    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:24 crc kubenswrapper[4824]: E0122 09:43:24.403894    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.420306    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.420431    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.420451    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.420482    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.420499    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.523519    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.523591    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.523616    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.523646    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.523669    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.627860    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.627968    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.627991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.628036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.628147    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.730997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.731061    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.731088    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.731119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.731142    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.835216    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.835284    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.835302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.835333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.835349    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.938013    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.938060    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.938072    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.938092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:24 crc kubenswrapper[4824]: I0122 09:43:24.938106    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:24Z","lastTransitionTime":"2026-01-22T09:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.041634    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.041699    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.041715    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.041742    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.041760    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.145392    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.145443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.145456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.145480    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.145495    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.248971    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.249027    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.249036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.249057    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.249069    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.341185    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"]
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.342761    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.346377    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.346523    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.351751    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.351817    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.351844    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.351882    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.351906    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.368520    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.370584    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 17:36:24.576703121 +0000 UTC
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.385113    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.391977    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.392038    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.392058    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.392075    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brwsj\" (UniqueName: \"kubernetes.io/projected/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-kube-api-access-brwsj\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.401915    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.416771    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.431291    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.455310    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.455388    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.455404    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.455427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.455447    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.456761    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.478781    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.493637    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.493722    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.493760    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.493794    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brwsj\" (UniqueName: \"kubernetes.io/projected/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-kube-api-access-brwsj\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.494602    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.494754    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.496261    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.499286    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.514542    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.517080    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brwsj\" (UniqueName: \"kubernetes.io/projected/606a4ae4-6d9d-46e1-a1c3-583b4737af6b-kube-api-access-brwsj\") pod \"ovnkube-control-plane-749d76644c-9948k\" (UID: \"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.532034    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.546780    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.557125    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.558323    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.558422    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.558448    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.558477    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.558499    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.569454    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.583556    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.594469    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.659855    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.662345    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.662432    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.662448    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.662474    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.662492    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: W0122 09:43:25.684392    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod606a4ae4_6d9d_46e1_a1c3_583b4737af6b.slice/crio-01c7a45c56441e0aca3c861b79d618d556adab1dc55ecea57e0702602c1cd34b WatchSource:0}: Error finding container 01c7a45c56441e0aca3c861b79d618d556adab1dc55ecea57e0702602c1cd34b: Status 404 returned error can't find the container with id 01c7a45c56441e0aca3c861b79d618d556adab1dc55ecea57e0702602c1cd34b
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.720651    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" event={"ID":"606a4ae4-6d9d-46e1-a1c3-583b4737af6b","Type":"ContainerStarted","Data":"01c7a45c56441e0aca3c861b79d618d556adab1dc55ecea57e0702602c1cd34b"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.725237    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/0.log"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.730156    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd" exitCode=1
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.730213    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.731121    4824 scope.go:117] "RemoveContainer" containerID="1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.743088    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.764861    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.764997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.765041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.765056    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.765077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.765271    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.781114    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.799230    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.813524    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.825968    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.840090    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.855961    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.867465    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.867486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.867493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.867507    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.867516    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.876629    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.887221    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.898472    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.911545    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.927705    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.943133    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.956581    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:25Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.970628    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.970673    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.970683    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.970703    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:25 crc kubenswrapper[4824]: I0122 09:43:25.970715    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:25Z","lastTransitionTime":"2026-01-22T09:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.073462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.073509    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.073519    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.073536    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.073549    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.177299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.177390    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.177408    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.177435    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.177453    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.279340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.279428    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.279441    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.279460    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.279474    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.370959    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 12:47:42.891251533 +0000 UTC
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.382480    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.382516    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.382528    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.382547    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.382563    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.403982    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.403981    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.404021    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:26 crc kubenswrapper[4824]: E0122 09:43:26.404211    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:26 crc kubenswrapper[4824]: E0122 09:43:26.404290    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:26 crc kubenswrapper[4824]: E0122 09:43:26.404401    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.485906    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.485949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.485958    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.485985    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.485994    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.589786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.589856    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.589873    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.589899    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.589917    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.693052    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.693108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.693127    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.693154    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.693171    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.796962    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.797016    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.797032    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.797056    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.797074    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.845220    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xxtjz"]
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.846127    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:26 crc kubenswrapper[4824]: E0122 09:43:26.846262    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.865066    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.890282    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.900709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.900752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.900769    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.900792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.900810    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:26Z","lastTransitionTime":"2026-01-22T09:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.911494    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjtgq\" (UniqueName: \"kubernetes.io/projected/f43b4618-291a-40d3-b1ca-32abfadaf376-kube-api-access-jjtgq\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.911838    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.912261    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.929432    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.948022    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.968745    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:26 crc kubenswrapper[4824]: I0122 09:43:26.990076    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:26Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.004133    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.004211    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.004235    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.004264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.004282    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.012797    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.012879    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjtgq\" (UniqueName: \"kubernetes.io/projected/f43b4618-291a-40d3-b1ca-32abfadaf376-kube-api-access-jjtgq\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:27 crc kubenswrapper[4824]: E0122 09:43:27.013047    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:27 crc kubenswrapper[4824]: E0122 09:43:27.013207    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:27.513167729 +0000 UTC m=+40.416644892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.020156    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.035859    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjtgq\" (UniqueName: \"kubernetes.io/projected/f43b4618-291a-40d3-b1ca-32abfadaf376-kube-api-access-jjtgq\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.050094    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.066246    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.081204    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.094787    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.107037    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.107092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.107110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.107133    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.107150    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.112638    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.130577    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.143997    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.155827    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.210429    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.210732    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.210760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.210795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.210818    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.314894    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.314964    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.314981    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.315041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.315060    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.371811    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 17:54:46.130972022 +0000 UTC
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.417506    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.417544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.417554    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.417570    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.417580    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.429325    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.452172    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.476595    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.507759    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.517667    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:27 crc kubenswrapper[4824]: E0122 09:43:27.517952    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:27 crc kubenswrapper[4824]: E0122 09:43:27.518080    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:28.518050625 +0000 UTC m=+41.421527678 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.519957    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.520011    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.520028    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.520052    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.520070    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.524450    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.596083    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.620752    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.622114    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.622136    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.622144    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.622158    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.622168    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.635856    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.647698    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.663064    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.678886    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.692545    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.706623    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.724606    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.724680    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.724691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.724716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.724728    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.725239    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.740871    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.755725    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.827736    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.827787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.827800    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.827823    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.827835    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.931065    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.931110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.931119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.931136    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:27 crc kubenswrapper[4824]: I0122 09:43:27.931146    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:27Z","lastTransitionTime":"2026-01-22T09:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.033843    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.033912    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.033932    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.033960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.033983    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.124537    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.124697    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.124729    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.124852    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.124876    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.124983    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:43:44.124934539 +0000 UTC m=+57.028411542 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.125035    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:44.125019922 +0000 UTC m=+57.028496925 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.125056    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:44.125045023 +0000 UTC m=+57.028522156 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.136985    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.137054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.137076    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.137110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.137135    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.226141    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.226258    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226383    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226418    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226430    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226481    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226497    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:44.226476494 +0000 UTC m=+57.129953487 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226511    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226531    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.226657    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:44.226631638 +0000 UTC m=+57.130108671 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.239260    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.239290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.239298    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.239312    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.239323    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.294973    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.306476    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.317728    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.328449    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.339168    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.341927    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.341954    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.341963    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.341980    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.341990    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.353003    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.368609    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.372275    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:41:32.131492345 +0000 UTC
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.381089    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.399160    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.403201    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.403220    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.403201    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.403204    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.403366    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.403475    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.403653    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.403785    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.426650    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.439045    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.444154    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.444207    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.444221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.444241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.444254    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.452412    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.464956    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.477173    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.488486    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.500958    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.512600    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.530116    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.530263    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: E0122 09:43:28.530319    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:30.530301582 +0000 UTC m=+43.433778575 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.548886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.548936    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.548953    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.548973    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.548986    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.652344    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.652611    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.652625    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.652644    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.652658    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.740391    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/0.log"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.743066    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.743497    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.744778    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" event={"ID":"606a4ae4-6d9d-46e1-a1c3-583b4737af6b","Type":"ContainerStarted","Data":"14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.754981    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.755026    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.755038    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.755054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.755071    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.767697    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.781508    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.795956    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.810396    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.825817    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.842854    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.855181    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.857445    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.857486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.857509    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.857530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.857544    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.868598    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.880393    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.896925    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.913248    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.925047    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.937484    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.949889    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.960806    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.960948    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.960981    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.961000    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.961012    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:28Z","lastTransitionTime":"2026-01-22T09:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.965609    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:28 crc kubenswrapper[4824]: I0122 09:43:28.985141    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:28Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.064329    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.064441    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.064465    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.064493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.064512    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.167380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.167413    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.167421    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.167438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.167450    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.270535    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.270584    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.270593    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.270611    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.270623    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.372406    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 00:28:29.068440386 +0000 UTC
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.372915    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.372950    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.372962    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.372980    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.372995    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.475126    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.475168    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.475181    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.475199    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.475212    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.578258    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.578315    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.578326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.578344    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.578371    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.681784    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.681833    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.681845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.681863    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.681874    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.749211    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" event={"ID":"606a4ae4-6d9d-46e1-a1c3-583b4737af6b","Type":"ContainerStarted","Data":"eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.751718    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/1.log"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.752656    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/0.log"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.762051    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785" exitCode=1
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.762436    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.762626    4824 scope.go:117] "RemoveContainer" containerID="1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.763210    4824 scope.go:117] "RemoveContainer" containerID="e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785"
Jan 22 09:43:29 crc kubenswrapper[4824]: E0122 09:43:29.763520    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.770424    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.784678    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.784724    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.784733    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.784751    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.784761    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.786993    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.802779    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.819267    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.832664    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.846295    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.857331    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.873176    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.886961    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.888224    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.888376    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.888470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.888557    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.888633    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.907413    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.925296    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.940822    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.957371    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.980090    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.991198    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.991490    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.991578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.991697    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.991781    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:29Z","lastTransitionTime":"2026-01-22T09:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:29 crc kubenswrapper[4824]: I0122 09:43:29.997954    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:29Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.010436    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.023273    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.039245    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.050941    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.063138    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.073791    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.086040    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.094535    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.094769    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.094882    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.094982    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.095079    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.102294    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.121315    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de6e0551d71a9e750a0c891dca15ca38f086e07acc189e74534900cfbc7e6dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:24Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268332    6093 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0122 09:43:23.268350    6093 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0122 09:43:23.268403    6093 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0122 09:43:23.268408    6093 handler.go:208] Removed *v1.Node event handler 2\\\\nI0122 09:43:23.268421    6093 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0122 09:43:23.268428    6093 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0122 09:43:23.268452    6093 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0122 09:43:23.268453    6093 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0122 09:43:23.268461    6093 handler.go:208] Removed *v1.Node event handler 7\\\\nI0122 09:43:23.268469    6093 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0122 09:43:23.268477    6093 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0122 09:43:23.268505    6093 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0122 09:43:23.269016    6093 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"message\\\":\\\"_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713245    6247 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713036    6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 09:43:29.713375    6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.132900    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.148731    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.160815    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.173811    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.184855    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.196963    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.197343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.197412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.197424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.197440    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.197449    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.208500    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.217467    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.300577    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.300627    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.300639    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.300656    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.300673    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.372915    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 08:37:23.084443334 +0000 UTC
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403248    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403291    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403287    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403259    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.403435    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403449    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403489    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403501    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.403531    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.403528    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.403702    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.403814    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.507557    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.507649    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.508057    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.508146    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.508171    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.549906    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.550118    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.550235    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:34.550206208 +0000 UTC m=+47.453683241 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.612471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.612538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.612556    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.612583    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.612603    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.716252    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.716316    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.716333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.716394    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.716412    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.770262    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/1.log"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.776718    4824 scope.go:117] "RemoveContainer" containerID="e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785"
Jan 22 09:43:30 crc kubenswrapper[4824]: E0122 09:43:30.776970    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.794532    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.814680    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.820264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.820302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.820314    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.820332    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.820345    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.835774    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.855679    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.875184    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.907515    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"message\\\":\\\"_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713245    6247 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713036    6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 09:43:29.713375    6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.923297    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.923349    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.923378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.923399    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.923411    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:30Z","lastTransitionTime":"2026-01-22T09:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.925456    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.943988    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.963529    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:30 crc kubenswrapper[4824]: I0122 09:43:30.987071    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:30Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.006265    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.022213    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.027452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.027510    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.027536    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.027568    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.027592    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.037672    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.055771    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.075120    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.091877    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:31Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.131325    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.131430    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.131449    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.131476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.131499    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.234292    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.234342    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.234371    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.234390    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.234400    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.337425    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.337489    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.337504    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.337531    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.337547    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.373099    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 13:24:42.65172829 +0000 UTC
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.439456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.439716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.439919    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.440039    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.440405    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.543101    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.543149    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.543163    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.543186    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.543205    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.646456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.646786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.646930    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.647072    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.647203    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.749719    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.749812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.749838    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.749882    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.749897    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.853179    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.853567    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.853780    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.854422    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.854468    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.957891    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.958320    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.958574    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.958800    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:31 crc kubenswrapper[4824]: I0122 09:43:31.958958    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:31Z","lastTransitionTime":"2026-01-22T09:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.062500    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.062550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.062560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.062578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.062588    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.165517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.165564    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.165577    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.165595    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.165608    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.268400    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.268488    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.268514    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.268547    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.268577    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.302333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.302475    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.302501    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.302533    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.302556    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.325515    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:32Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.332199    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.332259    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.332282    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.332314    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.332337    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.355607    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:32Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.360679    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.360808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.360886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.360980    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.361052    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.373548    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 23:57:57.297231451 +0000 UTC
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.381254    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:32Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.386558    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.386587    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.386597    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.386613    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.386623    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.401076    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:32Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.403265    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.403279    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.403393    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.403405    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.403506    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.403520    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.403719    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.403845    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.405486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.405605    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.405684    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.405758    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.405839    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.428892    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:32Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:32 crc kubenswrapper[4824]: E0122 09:43:32.429058    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.431213    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.431250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.431265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.431286    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.431300    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.535006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.535068    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.535089    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.535112    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.535128    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.638467    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.638548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.638567    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.638596    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.638616    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.742802    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.743168    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.743333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.743752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.744100    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.847446    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.847541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.847560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.847586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.847605    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.951187    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.951277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.951301    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.951333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:32 crc kubenswrapper[4824]: I0122 09:43:32.951417    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:32Z","lastTransitionTime":"2026-01-22T09:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.055856    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.056234    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.056426    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.056573    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.056698    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.160433    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.161189    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.161936    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.162146    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.162446    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.266054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.266548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.266830    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.267019    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.267153    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.370976    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.371090    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.371117    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.371146    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.371168    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.374225    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 20:43:16.88795778 +0000 UTC
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.480340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.482484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.482535    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.482564    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.482584    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.585948    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.585996    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.586007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.586026    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.586038    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.689105    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.689160    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.689177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.689200    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.689215    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.792210    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.792284    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.792302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.792327    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.792344    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.896109    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.896175    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.896189    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.896215    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:33 crc kubenswrapper[4824]: I0122 09:43:33.896232    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:33Z","lastTransitionTime":"2026-01-22T09:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.001056    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.001484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.001542    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.001585    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.001606    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.105464    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.105531    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.105544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.105564    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.105579    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.209036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.209102    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.209123    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.209152    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.209172    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.312476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.312524    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.312549    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.312572    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.312586    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.374785    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 08:20:27.303198259 +0000 UTC
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.403470    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.403529    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.403470    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.403739    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:34 crc kubenswrapper[4824]: E0122 09:43:34.403720    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:34 crc kubenswrapper[4824]: E0122 09:43:34.403814    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:34 crc kubenswrapper[4824]: E0122 09:43:34.403957    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:34 crc kubenswrapper[4824]: E0122 09:43:34.404124    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.416648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.416758    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.416780    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.416810    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.416836    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.520798    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.520865    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.520887    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.520919    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.520941    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.599713    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:34 crc kubenswrapper[4824]: E0122 09:43:34.599939    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:34 crc kubenswrapper[4824]: E0122 09:43:34.600071    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:42.600036479 +0000 UTC m=+55.503513512 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.624471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.624644    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.624670    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.624699    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.624726    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.727895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.727939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.727954    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.727973    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.727988    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.831341    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.831411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.831421    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.831439    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.831454    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.934728    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.934802    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.934819    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.934842    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:34 crc kubenswrapper[4824]: I0122 09:43:34.934860    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:34Z","lastTransitionTime":"2026-01-22T09:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.038691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.038760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.038781    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.038808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.038827    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.141275    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.141324    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.141340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.141403    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.141421    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.244898    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.244974    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.244997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.245028    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.245095    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.349192    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.349244    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.349258    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.349280    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.349293    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.375059    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 01:08:26.458171026 +0000 UTC
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.451844    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.451899    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.451915    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.451939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.451956    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.555301    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.555337    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.555378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.555399    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.555414    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.659661    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.659713    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.659730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.659756    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.659774    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.763614    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.763686    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.763705    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.763736    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.763758    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.867159    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.867211    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.867221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.867239    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.867248    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.969770    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.969860    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.969882    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.969939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:35 crc kubenswrapper[4824]: I0122 09:43:35.969958    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:35Z","lastTransitionTime":"2026-01-22T09:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.009557    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.027499    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"]
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.033337    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.052892    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.072549    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.073723    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.073807    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.073822    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.073843    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.073859    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.090303    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.110377    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.128417    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.146781    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.187698    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.189445    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.189495    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.189511    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.189535    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.189551    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.231202    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"message\\\":\\\"_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713245    6247 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713036    6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 09:43:29.713375    6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.244560    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.258967    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.271321    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.287752    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.291940    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.291992    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.292006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.292025    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.292037    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.304764    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.320770    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.336545    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:36Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.375850    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 00:37:12.099413446 +0000 UTC
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.394797    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.394855    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.394875    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.394896    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.394910    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.403306    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.403377    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.403499    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:36 crc kubenswrapper[4824]: E0122 09:43:36.403579    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.403652    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:36 crc kubenswrapper[4824]: E0122 09:43:36.403940    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:36 crc kubenswrapper[4824]: E0122 09:43:36.404100    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:36 crc kubenswrapper[4824]: E0122 09:43:36.404224    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.499058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.499138    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.499156    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.499185    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.499208    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.602880    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.602943    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.602960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.602982    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.603000    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.706520    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.706596    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.706618    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.706643    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.706670    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.809808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.809862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.809875    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.809895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.809907    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.913068    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.913110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.913122    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.913139    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:36 crc kubenswrapper[4824]: I0122 09:43:36.913152    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:36Z","lastTransitionTime":"2026-01-22T09:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.016978    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.017034    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.017049    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.017075    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.017091    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.120091    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.120134    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.120142    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.120160    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.120172    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.223728    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.223797    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.223821    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.223858    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.223883    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.327533    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.327876    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.327897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.327921    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.327939    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.377624    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 11:58:41.134063916 +0000 UTC
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.426098    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.431405    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.431454    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.431464    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.431484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.431497    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.440843    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.458418    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.472476    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.494489    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.515994    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.529654    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.535071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.535122    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.535136    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.535159    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.535173    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.551688    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.573106    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"message\\\":\\\"_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713245    6247 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713036    6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 09:43:29.713375    6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.586626    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.600822    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.617379    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.633247    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.638013    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.638076    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.638087    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.638109    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.638122    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.646703    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.659041    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.675226    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.686926    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.741277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.741330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.741341    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.741379    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.741397    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.844523    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.845068    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.845238    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.845461    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.845631    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.949204    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.949261    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.949277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.949304    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:37 crc kubenswrapper[4824]: I0122 09:43:37.949322    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:37Z","lastTransitionTime":"2026-01-22T09:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.052567    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.052789    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.052808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.052832    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.052845    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.156144    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.156221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.156234    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.156254    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.156267    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.259773    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.259829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.259843    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.259867    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.259879    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.363846    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.363898    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.363912    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.363929    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.363942    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.378651    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 11:16:53.964309309 +0000 UTC
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.403185    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.403265    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:38 crc kubenswrapper[4824]: E0122 09:43:38.403330    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.403186    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:38 crc kubenswrapper[4824]: E0122 09:43:38.403460    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.403524    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:38 crc kubenswrapper[4824]: E0122 09:43:38.403598    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:38 crc kubenswrapper[4824]: E0122 09:43:38.403913    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.467047    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.467118    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.467140    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.467174    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.467199    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.575712    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.575780    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.575798    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.575823    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.575839    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.678987    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.679243    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.679256    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.679276    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.679291    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.783402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.783451    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.783464    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.783483    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.783495    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.888406    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.888476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.888492    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.888514    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.888530    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.990752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.990789    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.990798    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.990817    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:38 crc kubenswrapper[4824]: I0122 09:43:38.990829    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:38Z","lastTransitionTime":"2026-01-22T09:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.094850    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.095412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.095658    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.095826    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.095953    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.200225    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.200285    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.200297    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.200319    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.200331    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.303667    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.303740    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.303758    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.303783    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.303799    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.380126    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 10:30:04.646497706 +0000 UTC
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.407777    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.407820    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.407836    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.407853    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.407866    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.511723    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.511767    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.511776    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.511792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.511805    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.614696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.615092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.615526    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.615691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.615812    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.719512    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.719952    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.720037    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.720122    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.720211    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.823031    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.823099    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.823121    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.823148    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.823169    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.927071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.927164    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.927187    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.927220    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:39 crc kubenswrapper[4824]: I0122 09:43:39.927247    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:39Z","lastTransitionTime":"2026-01-22T09:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.031439    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.031511    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.031539    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.031571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.031595    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.135946    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.136018    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.136044    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.136073    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.136091    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.239294    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.239340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.239387    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.239461    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.239480    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.342469    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.342520    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.342538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.342600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.342622    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.380789    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 05:54:05.560069483 +0000 UTC
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.403138    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.403194    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.403245    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:40 crc kubenswrapper[4824]: E0122 09:43:40.403322    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.403138    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:40 crc kubenswrapper[4824]: E0122 09:43:40.403605    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:40 crc kubenswrapper[4824]: E0122 09:43:40.403701    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:40 crc kubenswrapper[4824]: E0122 09:43:40.403786    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.445931    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.445988    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.445998    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.446015    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.446026    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.550158    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.550255    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.550285    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.550324    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.550349    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.653793    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.653837    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.653845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.653861    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.653870    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.756825    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.756872    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.756881    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.756896    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.756908    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.859775    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.860245    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.860497    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.860715    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.860937    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.963812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.964188    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.964287    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.964409    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:40 crc kubenswrapper[4824]: I0122 09:43:40.964502    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:40Z","lastTransitionTime":"2026-01-22T09:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.068599    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.068951    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.069098    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.069269    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.069448    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.173183    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.173224    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.173232    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.173247    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.173256    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.276104    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.276427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.276510    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.276586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.276711    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.380007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.380057    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.380069    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.380086    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.380099    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.381301    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 22:52:04.12000729 +0000 UTC
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.482901    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.482953    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.482969    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.482991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.483005    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.585702    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.585739    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.585747    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.585762    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.585773    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.689606    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.689673    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.689691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.689717    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.689734    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.792529    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.792604    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.792625    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.792662    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.792686    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.896265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.896337    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.896368    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.896392    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.896411    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.999277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.999385    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.999395    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.999411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:41 crc kubenswrapper[4824]: I0122 09:43:41.999420    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:41Z","lastTransitionTime":"2026-01-22T09:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.102380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.102413    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.102426    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.102443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.102454    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.205676    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.206019    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.206101    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.206191    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.206289    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.309419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.309468    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.309484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.309505    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.309520    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.382518    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 07:31:04.482103477 +0000 UTC
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.403132    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.403224    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.403272    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.403394    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.403974    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.404207    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.404303    4824 scope.go:117] "RemoveContainer" containerID="e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785"
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.404390    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.404461    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.413449    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.413482    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.413495    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.413514    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.413527    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.517417    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.517950    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.518178    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.518380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.518569    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.621755    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.621804    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.621819    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.621839    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.621853    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.695844    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.696005    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.696083    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:43:58.696061744 +0000 UTC m=+71.599538747 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.724262    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.724618    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.724815    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.725027    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.725184    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.786887    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.786943    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.786952    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.786968    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.786982    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.801794    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:42Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.806158    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.806197    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.806206    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.806226    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.806236    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.822391    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:42Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.829937    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.830003    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.830021    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.830049    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.830075    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.854570    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:42Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.859330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.859490    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.859627    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.859718    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.859815    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.875303    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:42Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.880414    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.880572    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.880665    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.880761    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.880851    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.896572    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:42Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:42 crc kubenswrapper[4824]: E0122 09:43:42.896705    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.899858    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.899894    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.899907    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.899925    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:42 crc kubenswrapper[4824]: I0122 09:43:42.899937    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:42Z","lastTransitionTime":"2026-01-22T09:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.003918    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.003978    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.003988    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.004027    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.004042    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.106965    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.107658    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.107696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.107727    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.107746    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.210666    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.210696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.210703    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.210718    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.210731    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.315101    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.315165    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.315186    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.315212    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.315234    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.383220    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 07:25:09.393396681 +0000 UTC
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.451869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.451935    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.451947    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.451966    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.451977    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.554997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.555047    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.555064    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.555086    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.555104    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.658144    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.658206    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.658216    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.658233    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.658243    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.761443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.761520    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.761540    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.761565    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.761585    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.833455    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/1.log"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.836950    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.837744    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.860476    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.864124    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.864164    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.864176    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.864206    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.864221    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.879759    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.896824    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.921759    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.935203    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.948122    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.963964    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.966937    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.966994    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.967013    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.967040    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.967059    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:43Z","lastTransitionTime":"2026-01-22T09:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:43 crc kubenswrapper[4824]: I0122 09:43:43.989287    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:43Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.014044    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"message\\\":\\\"_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713245    6247 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713036    6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 09:43:29.713375    6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.027819    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.046120    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.062466    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.071195    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.071247    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.071259    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.071280    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.071293    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.087090    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.106866    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.117938    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.130188    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.141984    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.174297    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.174381    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.174395    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.174415    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.174427    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.212305    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.212573    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.212648    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.212593569 +0000 UTC m=+89.116070592 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.212746    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.212837    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.212813536 +0000 UTC m=+89.116290559 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.212828    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.213113    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.213264    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.213232778 +0000 UTC m=+89.116709791 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.278571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.278630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.278642    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.278661    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.278680    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.314840    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.314971    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315275    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315344    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315434    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315275    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315501    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315527    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315536    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.315503329 +0000 UTC m=+89.218980362 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.315606    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:16.315576692 +0000 UTC m=+89.219053735 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.383418    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.383500    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.383519    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.383550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.383571    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.384451    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 19:56:18.24566058 +0000 UTC
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.404184    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.404248    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.404514    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.404572    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.405028    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.405190    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.405264    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.405460    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.486799    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.486871    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.486889    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.486918    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.486945    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.590672    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.591026    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.591097    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.591160    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.591217    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.695397    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.695465    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.695490    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.695524    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.695551    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.799667    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.799745    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.799763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.799789    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.799807    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.845053    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/2.log"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.846286    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/1.log"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.851774    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373" exitCode=1
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.851827    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.851879    4824 scope.go:117] "RemoveContainer" containerID="e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.853267    4824 scope.go:117] "RemoveContainer" containerID="3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"
Jan 22 09:43:44 crc kubenswrapper[4824]: E0122 09:43:44.853804    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.878926    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.900248    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.903491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.903573    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.903594    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.903624    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.903645    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:44Z","lastTransitionTime":"2026-01-22T09:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.934447    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.952344    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.977911    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:44 crc kubenswrapper[4824]: I0122 09:43:44.996501    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:44Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.006792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.006845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.006858    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.006909    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.006921    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.017312    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.034419    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.052248    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.081922    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9a00e898ee5dfa3a9600525ec130516c4993a102fa5568fb84aefb32cfad785\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"message\\\":\\\"_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713245    6247 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-config-operator/machine-config-operator]} name:Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.183:9001:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5b85277d-d9b7-4a68-8e4e-2b80594d9347}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:29.713036    6247 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0122 09:43:29.713375    6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.093384    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.103892    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.108742    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.108786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.108798    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.108820    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.108834    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.117859    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.132525    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.145740    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.157305    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.169614    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.211678    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.212005    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.212014    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.212031    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.212046    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.315611    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.315664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.315681    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.315705    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.315717    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.385262    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 22:00:38.834939986 +0000 UTC
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.418642    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.418686    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.418699    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.418720    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.418734    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.521265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.521332    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.521347    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.521393    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.521411    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.624830    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.624924    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.624951    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.624985    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.625009    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.727752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.727818    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.727836    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.727862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.727880    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.830963    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.831030    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.831052    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.831079    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.831100    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.858978    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/2.log"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.864641    4824 scope.go:117] "RemoveContainer" containerID="3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"
Jan 22 09:43:45 crc kubenswrapper[4824]: E0122 09:43:45.864889    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.878903    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.895276    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.917014    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.935854    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.935894    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.935903    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.935919    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.935930    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:45Z","lastTransitionTime":"2026-01-22T09:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.940335    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.960599    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:45 crc kubenswrapper[4824]: I0122 09:43:45.978846    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:45Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.002818    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.019174    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.034561    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.039273    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.039302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.039317    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.039336    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.039346    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.049261    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.064527    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.076532    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.088337    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.104879    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.120166    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.139415    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.142650    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.142700    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.142712    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.142731    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.142745    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.157416    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:46Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.245812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.246081    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.246229    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.246316    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.246434    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.349518    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.349559    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.349571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.349587    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.349598    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.386039    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 18:40:11.295777384 +0000 UTC
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.403345    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.403455    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:46 crc kubenswrapper[4824]: E0122 09:43:46.403534    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.403601    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:46 crc kubenswrapper[4824]: E0122 09:43:46.403738    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:46 crc kubenswrapper[4824]: E0122 09:43:46.403994    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.404441    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:46 crc kubenswrapper[4824]: E0122 09:43:46.404603    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.452582    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.452674    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.452696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.452731    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.452755    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.555688    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.555738    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.555751    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.555771    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.555784    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.658478    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.658515    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.658523    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.658538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.658549    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.761724    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.761785    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.761802    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.761827    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.761847    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.864973    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.865062    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.865084    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.865112    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.865133    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.968212    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.968277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.968294    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.968319    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:46 crc kubenswrapper[4824]: I0122 09:43:46.968337    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:46Z","lastTransitionTime":"2026-01-22T09:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.072759    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.072849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.072869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.072895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.072913    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.175805    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.176080    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.176157    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.176241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.176317    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.279307    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.279413    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.279431    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.279453    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.279465    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.382960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.383048    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.383076    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.383113    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.383137    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.386807    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 16:37:29.015075494 +0000 UTC
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.422165    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.445556    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.480185    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.485888    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.485930    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.485940    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.485960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.485973    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.498064    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.516878    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.531148    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.547645    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.562278    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.581224    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.589525    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.589569    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.589582    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.589603    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.589616    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.598856    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.613534    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.630483    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.644425    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.658816    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.673019    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.687930    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.691830    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.691904    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.691927    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.691954    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.691981    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.703773    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:47Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.795888    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.796262    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.796378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.796476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.796553    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.899327    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.899425    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.899441    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.899464    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:47 crc kubenswrapper[4824]: I0122 09:43:47.899478    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:47Z","lastTransitionTime":"2026-01-22T09:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.001945    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.001994    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.002005    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.002024    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.002034    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.105564    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.105610    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.105620    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.105639    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.105651    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.208035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.208125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.208143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.208170    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.208187    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.310939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.311001    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.311015    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.311035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.311045    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.388070    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 04:45:34.467906522 +0000 UTC
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.403609    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.403659    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.403688    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:48 crc kubenswrapper[4824]: E0122 09:43:48.404461    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.403769    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:48 crc kubenswrapper[4824]: E0122 09:43:48.404304    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:48 crc kubenswrapper[4824]: E0122 09:43:48.404091    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:48 crc kubenswrapper[4824]: E0122 09:43:48.404943    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.413486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.413532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.413550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.413572    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.413587    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.517419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.517493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.517515    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.517543    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.517560    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.620284    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.620554    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.620631    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.620706    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.620879    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.723628    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.723946    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.724041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.724125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.724216    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.827949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.828281    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.828470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.828690    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.828781    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.932164    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.932233    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.932254    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.932281    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:48 crc kubenswrapper[4824]: I0122 09:43:48.932303    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:48Z","lastTransitionTime":"2026-01-22T09:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.035696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.035763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.035780    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.035803    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.035820    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.138921    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.139008    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.139031    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.139060    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.139159    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.242948    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.243002    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.243014    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.243036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.243050    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.346100    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.346147    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.346161    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.346183    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.346200    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.389700    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 15:14:05.18811437 +0000 UTC
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.448315    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.448606    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.448727    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.448855    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.448975    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.552124    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.552162    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.552171    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.552187    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.552200    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.655053    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.655083    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.655091    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.655107    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.655117    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.757891    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.757921    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.757929    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.757944    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.757953    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.860548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.860622    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.860639    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.860664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.860681    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.962459    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.962740    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.962800    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.962859    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:49 crc kubenswrapper[4824]: I0122 09:43:49.962973    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:49Z","lastTransitionTime":"2026-01-22T09:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.065077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.065302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.065403    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.065501    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.065597    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.168438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.168482    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.168493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.168508    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.168519    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.271191    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.271223    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.271231    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.271245    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.271254    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.373534    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.373891    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.374038    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.374178    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.374292    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.390225    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 09:31:33.257440067 +0000 UTC
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.403667    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.403730    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.403778    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.403811    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:50 crc kubenswrapper[4824]: E0122 09:43:50.403852    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:50 crc kubenswrapper[4824]: E0122 09:43:50.403945    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:50 crc kubenswrapper[4824]: E0122 09:43:50.404010    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:50 crc kubenswrapper[4824]: E0122 09:43:50.404164    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.477402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.477486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.477506    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.477536    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.477556    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.580557    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.580590    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.580600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.580616    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.580640    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.684203    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.684252    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.684268    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.684287    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.684302    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.787418    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.787467    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.787484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.787525    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.787537    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.889891    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.889956    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.889967    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.889989    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.890002    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.993385    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.993434    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.993446    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.993467    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:50 crc kubenswrapper[4824]: I0122 09:43:50.993484    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:50Z","lastTransitionTime":"2026-01-22T09:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.095774    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.095812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.095822    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.095839    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.095850    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.198462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.198514    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.198524    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.198541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.198550    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.380644    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.380694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.380709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.380728    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.380738    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.390924    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 21:09:11.801767066 +0000 UTC
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.483311    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.483386    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.483397    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.483417    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.483427    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.586426    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.586494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.586521    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.586553    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.586575    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.689743    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.689792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.689802    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.689820    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.689830    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.792592    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.792626    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.792634    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.792648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.792658    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.895401    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.895448    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.895458    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.895475    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.895485    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.998863    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.998905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.998914    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.998931    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:51 crc kubenswrapper[4824]: I0122 09:43:51.998941    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:51Z","lastTransitionTime":"2026-01-22T09:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.102607    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.102664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.102672    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.102687    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.102697    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.205672    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.205711    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.205721    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.205735    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.205747    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.309004    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.309072    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.309092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.309123    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.309145    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.391045    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 23:37:46.553566106 +0000 UTC
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.403159    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.403210    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.403305    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.403188    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.403584    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.403488    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.403787    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.403827    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.411303    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.411468    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.411549    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.411630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.411719    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.513539    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.513604    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.513621    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.513651    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.513669    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.615918    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.615972    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.615990    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.616015    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.616033    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.718612    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.720444    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.720479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.720508    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.720525    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.822886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.822926    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.822934    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.822949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.822959    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.925654    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.925689    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.925698    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.925714    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.925724    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.928730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.928754    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.928763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.928773    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.928781    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.940831    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:52Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.944762    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.944803    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.944824    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.944848    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.944864    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.956084    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:52Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.960752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.960807    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.960820    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.960841    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.960857    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.979100    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:52Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.982737    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.982773    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.982809    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.982830    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:52 crc kubenswrapper[4824]: I0122 09:43:52.982843    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:52Z","lastTransitionTime":"2026-01-22T09:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:52 crc kubenswrapper[4824]: E0122 09:43:52.998677    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:52Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.002774    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.002950    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.003073    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.003210    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.003337    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: E0122 09:43:53.018089    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:53Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:53 crc kubenswrapper[4824]: E0122 09:43:53.018246    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.027909    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.027939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.027951    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.027968    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.027980    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.130883    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.130953    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.130971    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.131001    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.131022    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.234006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.234075    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.234089    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.234109    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.234122    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.336478    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.336529    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.336540    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.336562    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.336572    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.392323    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 10:39:16.696663443 +0000 UTC
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.439452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.439531    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.439546    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.439567    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.439581    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.542980    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.543035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.543048    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.543066    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.543077    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.645427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.645465    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.645474    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.645496    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.645505    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.747700    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.747741    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.747752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.747769    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.747780    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.850140    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.850219    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.850233    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.850254    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.850268    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.952653    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.952707    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.952721    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.952744    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:53 crc kubenswrapper[4824]: I0122 09:43:53.952757    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:53Z","lastTransitionTime":"2026-01-22T09:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.055462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.055526    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.055544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.055569    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.055593    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.158321    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.158453    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.158479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.158513    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.158536    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.261783    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.261846    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.261856    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.261876    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.261888    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.364968    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.365045    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.365069    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.365098    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.365119    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.393592    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 07:20:20.82514957 +0000 UTC
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.403940    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.403941    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.404008    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.404045    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:54 crc kubenswrapper[4824]: E0122 09:43:54.404254    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:54 crc kubenswrapper[4824]: E0122 09:43:54.404468    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:54 crc kubenswrapper[4824]: E0122 09:43:54.404580    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:54 crc kubenswrapper[4824]: E0122 09:43:54.404665    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.468688    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.468764    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.468789    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.468823    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.468846    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.572427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.572469    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.572480    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.572500    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.572512    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.675563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.675619    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.675631    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.675651    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.675661    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.778572    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.778652    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.778669    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.778694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.778715    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.881455    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.881520    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.881542    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.881570    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.881589    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.985152    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.985201    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.985217    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.985241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:54 crc kubenswrapper[4824]: I0122 09:43:54.985258    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:54Z","lastTransitionTime":"2026-01-22T09:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.088228    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.088263    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.088274    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.088292    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.088304    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.192186    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.192253    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.192273    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.192306    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.192327    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.295203    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.295271    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.295288    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.295311    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.295328    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.394196    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 04:23:49.065098725 +0000 UTC
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.398495    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.398547    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.398561    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.398588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.398614    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.503250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.503319    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.503344    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.503412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.503438    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.606497    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.606534    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.606545    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.606563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.606575    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.709869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.709905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.709916    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.709934    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.709944    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.813908    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.813951    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.813962    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.813979    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.813993    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.917313    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.917386    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.917398    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.917417    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:55 crc kubenswrapper[4824]: I0122 09:43:55.917434    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:55Z","lastTransitionTime":"2026-01-22T09:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.019933    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.019973    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.019982    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.019997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.020007    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.123506    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.123550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.123559    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.123575    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.123586    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.225867    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.225914    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.225923    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.225942    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.225952    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.328726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.328782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.328795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.328845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.328868    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.395173    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 15:11:21.976040921 +0000 UTC
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.403674    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.403703    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.403714    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:56 crc kubenswrapper[4824]: E0122 09:43:56.403836    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.403674    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:56 crc kubenswrapper[4824]: E0122 09:43:56.403977    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:56 crc kubenswrapper[4824]: E0122 09:43:56.404042    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:56 crc kubenswrapper[4824]: E0122 09:43:56.404093    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.431566    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.431605    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.431615    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.431631    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.431645    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.533726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.533766    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.533777    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.533794    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.533805    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.636800    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.636837    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.636846    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.636862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.636872    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.740045    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.740108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.740127    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.740153    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.740170    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.843543    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.843598    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.843612    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.843631    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.843643    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.946641    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.946716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.946734    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.946762    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:56 crc kubenswrapper[4824]: I0122 09:43:56.946782    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:56Z","lastTransitionTime":"2026-01-22T09:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.049786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.049822    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.049833    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.049849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.049860    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.152868    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.152921    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.152933    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.152954    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.152966    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.256154    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.256197    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.256211    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.256229    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.256241    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.358950    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.359005    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.359014    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.359032    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.359042    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.395392    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 15:36:09.284033421 +0000 UTC
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.417698    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"]
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.421495    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.434268    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.460453    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.462056    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.462183    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.462298    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.462320    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.462330    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.482454    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.520576    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.533744    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.547008    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.560208    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.565268    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.565319    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.565330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.565351    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.565388    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.573542    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.587905    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.600216    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.621222    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.638716    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.656140    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.668844    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.668885    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.668897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.668917    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.668930    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.671832    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.692164    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.703609    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:43:57Z is after 2025-08-24T17:21:41Z"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.770780    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.771328    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.771425    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.771494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.771553    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.874476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.874526    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.874538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.874558    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.874571    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.976859    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.976884    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.976892    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.976906    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:57 crc kubenswrapper[4824]: I0122 09:43:57.976915    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:57Z","lastTransitionTime":"2026-01-22T09:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.078736    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.078772    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.078782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.078796    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.078806    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.181250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.181290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.181321    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.181338    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.181347    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.283661    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.283706    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.283716    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.283730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.283738    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.385896    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.385963    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.385983    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.386015    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.386038    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.396034    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 12:41:00.979344422 +0000 UTC
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.403330    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.403350    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.403434    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.403491    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:43:58 crc kubenswrapper[4824]: E0122 09:43:58.403580    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:43:58 crc kubenswrapper[4824]: E0122 09:43:58.403679    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:43:58 crc kubenswrapper[4824]: E0122 09:43:58.403789    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:43:58 crc kubenswrapper[4824]: E0122 09:43:58.403961    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.488452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.488499    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.488511    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.488527    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.488538    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.591438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.591517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.591530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.591549    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.591561    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.694107    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.694173    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.694184    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.694201    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.694211    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.787548    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:43:58 crc kubenswrapper[4824]: E0122 09:43:58.787771    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:58 crc kubenswrapper[4824]: E0122 09:43:58.788657    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:44:30.788625096 +0000 UTC m=+103.692102119 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.796411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.796451    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.796461    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.796479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.796491    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.899243    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.899608    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.899752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.899893    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:58 crc kubenswrapper[4824]: I0122 09:43:58.900024    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:58Z","lastTransitionTime":"2026-01-22T09:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.002960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.003048    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.003070    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.003101    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.003124    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.105550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.105615    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.105626    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.105647    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.105660    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.208278    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.208409    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.208437    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.208472    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.208495    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.316913    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.316969    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.316980    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.317001    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.317013    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.397004    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 18:02:23.163826138 +0000 UTC
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.419961    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.419990    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.420003    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.420017    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.420027    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.523831    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.524258    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.524277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.524301    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.524317    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.627529    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.627570    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.627582    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.627603    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.627613    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.730657    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.730723    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.730744    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.730768    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.730785    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.834195    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.834250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.834260    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.834280    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.834293    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.937489    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.937569    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.937594    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.937621    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:43:59 crc kubenswrapper[4824]: I0122 09:43:59.937643    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:43:59Z","lastTransitionTime":"2026-01-22T09:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.041047    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.041141    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.041159    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.041184    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.041205    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.143944    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.143998    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.144007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.144022    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.144032    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.246462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.246786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.246881    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.246996    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.247114    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.350853    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.350907    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.350925    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.350951    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.350968    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.397716    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 09:06:35.055445072 +0000 UTC
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.403567    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.403675    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:00 crc kubenswrapper[4824]: E0122 09:44:00.403710    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.403782    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.403935    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:00 crc kubenswrapper[4824]: E0122 09:44:00.403981    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:00 crc kubenswrapper[4824]: E0122 09:44:00.404056    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:00 crc kubenswrapper[4824]: E0122 09:44:00.404219    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.405082    4824 scope.go:117] "RemoveContainer" containerID="3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"
Jan 22 09:44:00 crc kubenswrapper[4824]: E0122 09:44:00.405300    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.453616    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.453657    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.453668    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.453687    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.453698    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.556283    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.556330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.556343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.556378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.556392    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.659846    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.659965    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.659991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.660019    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.660038    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.762703    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.762766    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.762789    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.762821    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.762847    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.866271    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.866486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.866529    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.866563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.866606    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.969904    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.969939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.969949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.969966    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:00 crc kubenswrapper[4824]: I0122 09:44:00.969980    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:00Z","lastTransitionTime":"2026-01-22T09:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.073221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.073262    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.073275    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.073293    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.073306    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.175970    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.176008    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.176017    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.176031    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.176040    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.278741    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.278849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.278866    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.278916    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.278937    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.382063    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.382118    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.382130    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.382152    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.382167    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.398562    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 19:07:57.163632977 +0000 UTC
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.484776    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.484829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.484840    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.484857    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.484868    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.494567    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/0.log"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.494624    4824 generic.go:334] "Generic (PLEG): container finished" podID="00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c" containerID="660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df" exitCode=1
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.494657    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerDied","Data":"660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.495042    4824 scope.go:117] "RemoveContainer" containerID="660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.510792    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.527410    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.543217    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.559908    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.591326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.591395    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.591411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.591435    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.591451    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.605173    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.618623    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.632429    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.645828    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.660851    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.678263    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.695484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.695533    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.695542    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.695557    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.695567    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.695599    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.707087    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.717896    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.731079    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.744924    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.757379    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.771763    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.784830    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:01Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.797631    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.797671    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.797683    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.797700    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.797709    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.901097    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.901159    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.901181    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.901209    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:01 crc kubenswrapper[4824]: I0122 09:44:01.901233    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:01Z","lastTransitionTime":"2026-01-22T09:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.003878    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.003959    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.003977    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.004001    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.004018    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.106659    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.106707    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.106722    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.106740    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.106751    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.209625    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.209665    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.209680    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.209698    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.209711    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.313431    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.313472    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.313480    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.313497    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.313507    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.398765    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 12:21:02.906062695 +0000 UTC
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.403464    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.403491    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.403573    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:02 crc kubenswrapper[4824]: E0122 09:44:02.403734    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.403769    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:02 crc kubenswrapper[4824]: E0122 09:44:02.403900    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:02 crc kubenswrapper[4824]: E0122 09:44:02.404051    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:02 crc kubenswrapper[4824]: E0122 09:44:02.404144    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.416497    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.416541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.416552    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.416565    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.416576    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.500921    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/0.log"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.501008    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerStarted","Data":"814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.517854    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.519285    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.519347    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.519377    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.519398    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.519411    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.534273    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.551270    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.569844    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.584321    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.603678    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.623473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.623537    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.623554    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.623578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.623598    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.627885    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.657085    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.670526    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.690405    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.710721    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.727499    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.727584    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.727603    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.727629    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.727648    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.730630    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.750681    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.769762    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.788510    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.805581    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.826225    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.830883    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.830943    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.830961    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.830987    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.831004    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.842326    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:02Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.933829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.934481    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.934512    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.934546    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:02 crc kubenswrapper[4824]: I0122 09:44:02.934565    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:02Z","lastTransitionTime":"2026-01-22T09:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.037812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.037879    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.037911    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.037941    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.037963    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.141221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.141315    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.141349    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.141424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.141446    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.245098    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.245173    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.245191    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.245218    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.245236    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.252221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.252310    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.252336    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.252402    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.252424    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: E0122 09:44:03.269449    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:03Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.274163    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.274236    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.274263    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.274290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.274307    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: E0122 09:44:03.287280    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:03Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.291497    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.291533    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.291542    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.291555    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.291567    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: E0122 09:44:03.308643    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:03Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.314079    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.314151    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.314175    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.314208    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.314231    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: E0122 09:44:03.333758    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:03Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.337735    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.337782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.337792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.337812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.337824    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: E0122 09:44:03.348338    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:03Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:03 crc kubenswrapper[4824]: E0122 09:44:03.348477    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.350261    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.350289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.350299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.350312    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.350321    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.399953    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 07:21:17.674724442 +0000 UTC
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.453292    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.453348    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.453405    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.453428    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.453447    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.556977    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.557067    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.557093    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.557125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.557150    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.660980    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.661068    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.661085    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.661149    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.661167    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.765030    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.765089    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.765103    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.765125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.765141    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.869323    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.869396    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.869414    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.869433    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.869444    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.973087    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.973470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.973500    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.973530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:03 crc kubenswrapper[4824]: I0122 09:44:03.973552    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:03Z","lastTransitionTime":"2026-01-22T09:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.077401    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.077472    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.077494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.077525    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.077549    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.181075    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.181149    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.181186    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.181220    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.181241    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.284150    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.284212    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.284234    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.284264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.284285    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.387782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.387851    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.387869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.387898    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.387916    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.401011    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 11:29:17.626120038 +0000 UTC
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.403410    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.403453    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.403508    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.403508    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:04 crc kubenswrapper[4824]: E0122 09:44:04.403567    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:04 crc kubenswrapper[4824]: E0122 09:44:04.403699    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:04 crc kubenswrapper[4824]: E0122 09:44:04.403810    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:04 crc kubenswrapper[4824]: E0122 09:44:04.404583    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.490861    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.491165    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.491335    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.491548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.491683    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.594458    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.594514    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.594535    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.594563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.594584    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.697149    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.697289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.697319    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.697387    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.697415    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.801522    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.801586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.801609    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.801638    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.801730    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.905343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.905452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.905470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.905495    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:04 crc kubenswrapper[4824]: I0122 09:44:04.905514    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:04Z","lastTransitionTime":"2026-01-22T09:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.008451    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.009487    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.009752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.009963    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.010159    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.115641    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.115749    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.115775    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.116340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.116604    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.220011    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.220049    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.220060    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.220077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.220088    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.322656    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.322894    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.322976    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.323097    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.323182    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.401957    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 01:35:54.639390553 +0000 UTC
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.425937    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.425978    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.425991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.426007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.426019    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.529151    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.529260    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.529286    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.529317    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.529339    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.632224    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.632306    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.632329    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.632389    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.632413    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.735265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.735348    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.735409    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.735443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.735471    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.839064    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.839137    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.839159    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.839189    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.839208    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.941864    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.941906    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.941916    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.941937    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:05 crc kubenswrapper[4824]: I0122 09:44:05.941949    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:05Z","lastTransitionTime":"2026-01-22T09:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.045142    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.045195    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.045207    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.045229    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.045246    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.148115    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.148175    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.148195    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.148220    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.148238    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.250807    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.250892    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.250903    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.250922    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.250935    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.353424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.353477    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.353494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.353516    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.353533    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.403386    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 07:03:46.550551518 +0000 UTC
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.403534    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.403598    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.403577    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.403570    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:06 crc kubenswrapper[4824]: E0122 09:44:06.403776    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:06 crc kubenswrapper[4824]: E0122 09:44:06.403881    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:06 crc kubenswrapper[4824]: E0122 09:44:06.403985    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:06 crc kubenswrapper[4824]: E0122 09:44:06.404144    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.456996    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.457065    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.457079    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.457105    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.457121    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.559523    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.559583    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.559600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.559629    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.559647    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.663630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.663692    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.663704    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.663722    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.663734    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.766777    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.766821    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.766832    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.766850    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.766864    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.870527    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.870578    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.870589    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.870608    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.870623    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.973237    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.973289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.973301    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.973323    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:06 crc kubenswrapper[4824]: I0122 09:44:06.973344    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:06Z","lastTransitionTime":"2026-01-22T09:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.076071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.076116    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.076126    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.076148    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.076162    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.180256    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.180310    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.180329    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.180390    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.180417    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.288300    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.288351    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.288378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.288395    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.288405    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.391089    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.391136    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.391146    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.391162    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.391172    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.403720    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:25:37.942553639 +0000 UTC
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.415681    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.429620    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.445047    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.459984    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.472947    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.486613    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.492940    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.492976    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.492987    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.493006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.493019    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.500150    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.512613    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.530203    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.546178    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.556233    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.568230    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.580456    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.593773    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.596155    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.596191    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.596202    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.596217    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.596226    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.608779    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.621135    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.640572    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.652594    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:07Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.699399    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.699458    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.699471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.699491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.699504    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.802622    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.802702    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.802725    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.802756    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.802778    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.905060    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.905371    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.905382    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.905401    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:07 crc kubenswrapper[4824]: I0122 09:44:07.905413    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:07Z","lastTransitionTime":"2026-01-22T09:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.007630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.007663    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.007675    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.007693    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.007703    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.110426    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.110470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.110479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.110494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.110503    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.213009    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.213062    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.213078    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.213097    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.213108    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.316901    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.317035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.317058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.317087    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.317110    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.403414    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.403414    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.403437    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.403538    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.404216    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 13:59:49.5655064 +0000 UTC
Jan 22 09:44:08 crc kubenswrapper[4824]: E0122 09:44:08.404250    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:08 crc kubenswrapper[4824]: E0122 09:44:08.404539    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:08 crc kubenswrapper[4824]: E0122 09:44:08.404615    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:08 crc kubenswrapper[4824]: E0122 09:44:08.404733    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.420210    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.420250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.420290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.420309    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.420324    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.523620    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.523682    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.523698    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.523724    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.523742    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.626596    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.626663    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.626683    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.626706    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.626724    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.729175    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.729248    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.729263    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.729285    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.729297    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.831734    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.831782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.831803    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.831826    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.831847    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.934401    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.934473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.934491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.934516    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:08 crc kubenswrapper[4824]: I0122 09:44:08.934538    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:08Z","lastTransitionTime":"2026-01-22T09:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.037393    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.037438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.037450    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.037473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.037484    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.140978    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.141037    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.141047    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.141065    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.141078    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.243636    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.243673    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.243680    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.243695    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.243705    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.345755    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.345795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.345803    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.345818    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.345828    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.404660    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 23:56:44.569132754 +0000 UTC
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.447624    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.447663    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.447671    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.447684    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.447693    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.550097    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.550138    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.550150    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.550167    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.550180    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.653309    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.653381    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.653399    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.653420    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.653435    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.756007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.756071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.756091    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.756112    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.756125    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.860226    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.860313    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.860338    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.860435    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.860463    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.963575    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.963640    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.963659    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.963686    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:09 crc kubenswrapper[4824]: I0122 09:44:09.963707    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:09Z","lastTransitionTime":"2026-01-22T09:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.065786    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.065854    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.065865    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.065883    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.065895    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.168016    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.168070    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.168087    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.168108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.168128    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.271623    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.271653    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.271664    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.271714    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.271726    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.375177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.375269    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.375293    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.375318    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.375334    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.403855    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:10 crc kubenswrapper[4824]: E0122 09:44:10.404023    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.404264    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:10 crc kubenswrapper[4824]: E0122 09:44:10.404347    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.404522    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.404663    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:10 crc kubenswrapper[4824]: E0122 09:44:10.404746    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:10 crc kubenswrapper[4824]: E0122 09:44:10.404825    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.404815    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 22:55:04.246874012 +0000 UTC
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.478264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.478325    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.478340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.478383    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.478399    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.581448    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.581508    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.581519    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.581542    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.581554    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.683986    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.684023    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.684035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.684053    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.684064    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.788202    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.788261    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.788279    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.788300    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.788315    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.890540    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.890590    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.890600    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.890617    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.890628    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.994230    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.994340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.994424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.994462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:10 crc kubenswrapper[4824]: I0122 09:44:10.994486    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:10Z","lastTransitionTime":"2026-01-22T09:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.097117    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.097194    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.097216    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.097248    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.097270    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.199395    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.199425    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.199432    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.199447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.199458    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.302182    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.302220    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.302232    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.302250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.302261    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.404901    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.404947    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.404961    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.404955    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 07:15:07.348681281 +0000 UTC
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.404982    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.405018    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.584030    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.584082    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.584093    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.584111    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.584124    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.687281    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.687337    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.687368    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.687394    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.687409    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.790257    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.790520    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.790597    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.790683    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.790798    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.893887    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.893934    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.893949    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.894004    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.894024    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.996702    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.997072    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.997258    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.997496    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:11 crc kubenswrapper[4824]: I0122 09:44:11.997718    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:11Z","lastTransitionTime":"2026-01-22T09:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.100220    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.100286    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.100298    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.100318    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.100332    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.208230    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.208270    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.208279    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.208299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.208310    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.311068    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.311149    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.311172    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.311201    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.311221    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.403619    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:12 crc kubenswrapper[4824]: E0122 09:44:12.404019    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.403651    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:12 crc kubenswrapper[4824]: E0122 09:44:12.404454    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.403630    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:12 crc kubenswrapper[4824]: E0122 09:44:12.404798    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.403741    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:12 crc kubenswrapper[4824]: E0122 09:44:12.404963    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.405044    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 22:54:47.002293519 +0000 UTC
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.413809    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.413846    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.413857    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.413876    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.413887    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.516327    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.516619    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.516712    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.516773    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.516832    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.624591    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.624623    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.624632    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.624648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.624657    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.727840    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.727927    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.727964    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.727997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.728020    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.831230    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.831637    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.831719    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.831810    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.831908    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.934276    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.934562    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.934647    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.934728    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:12 crc kubenswrapper[4824]: I0122 09:44:12.934806    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:12Z","lastTransitionTime":"2026-01-22T09:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.037935    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.037984    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.038002    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.038026    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.038045    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.140640    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.140698    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.140713    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.140771    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.140788    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.243027    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.243093    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.243117    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.243143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.243170    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.346289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.346399    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.346419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.346447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.346469    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.406521    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:36:56.798281859 +0000 UTC
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.449823    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.449890    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.449909    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.449932    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.449950    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.530020    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.530096    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.530121    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.530154    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.530180    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: E0122 09:44:13.551077    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.557438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.557584    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.557604    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.557630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.557652    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: E0122 09:44:13.576651    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.581676    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.581748    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.581771    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.582132    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.582446    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: E0122 09:44:13.598066    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.603317    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.603474    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.603566    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.603649    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.603686    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: E0122 09:44:13.618649    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.622730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.622781    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.622792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.622812    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.622826    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: E0122 09:44:13.638064    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:13Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:13 crc kubenswrapper[4824]: E0122 09:44:13.638230    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.641836    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.641878    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.641888    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.641904    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.641915    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.745185    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.745528    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.745610    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.745694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.745857    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.848544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.848595    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.848607    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.848625    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.848637    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.981405    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.981630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.981727    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.981797    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:13 crc kubenswrapper[4824]: I0122 09:44:13.981870    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:13Z","lastTransitionTime":"2026-01-22T09:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.083865    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.083894    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.083905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.083922    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.083933    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.186271    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.186322    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.186334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.186387    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.186413    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.288318    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.288378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.288396    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.288419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.288436    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.392739    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.392970    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.393054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.393132    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.393198    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.403349    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.403520    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.403470    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.403405    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:14 crc kubenswrapper[4824]: E0122 09:44:14.403777    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:14 crc kubenswrapper[4824]: E0122 09:44:14.403906    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:14 crc kubenswrapper[4824]: E0122 09:44:14.404020    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:14 crc kubenswrapper[4824]: E0122 09:44:14.404139    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.406628    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 08:47:15.834928666 +0000 UTC
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.495907    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.495941    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.495955    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.495973    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.495984    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.598538    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.598598    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.598617    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.598643    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.598661    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.701480    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.701541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.701566    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.701591    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.701608    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.805914    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.806019    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.806041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.806074    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.806098    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.909760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.909830    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.909847    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.909874    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:14 crc kubenswrapper[4824]: I0122 09:44:14.909891    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:14Z","lastTransitionTime":"2026-01-22T09:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.012837    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.012875    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.012884    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.012899    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.012909    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.115654    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.115705    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.115721    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.115740    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.115753    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.218898    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.218932    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.218941    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.218958    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.218969    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.321865    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.321904    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.321914    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.321932    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.321943    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.404923    4824 scope.go:117] "RemoveContainer" containerID="3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.406803    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 04:59:00.214164413 +0000 UTC
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.424513    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.424559    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.424569    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.424588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.424600    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.528159    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.528198    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.528210    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.528227    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.528237    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.631974    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.632011    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.632024    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.632043    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.632055    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.735249    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.735334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.735351    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.735416    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.735438    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.838795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.838862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.838872    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.838891    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.838903    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.942035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.942095    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.942113    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.942136    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:15 crc kubenswrapper[4824]: I0122 09:44:15.942153    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:15Z","lastTransitionTime":"2026-01-22T09:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.045658    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.045715    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.045726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.045744    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.045757    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.148495    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.148552    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.148566    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.148588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.148605    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.250917    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.250961    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.250974    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.250991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.251002    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.312116    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.312252    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.312281    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.312413    4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.312473    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.312457155 +0000 UTC m=+153.215934148 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.312644    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.312637651 +0000 UTC m=+153.216114644 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.312703    4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.312723    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.312718273 +0000 UTC m=+153.216195256 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.353190    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.353221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.353229    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.353243    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.353253    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.403104    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.403320    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.403650    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.403771    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.403971    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.404055    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.404240    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.404341    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.407767    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 05:28:53.993792818 +0000 UTC
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.413652    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.413739    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.413839    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.413868    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.413880    4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.413942    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.413924914 +0000 UTC m=+153.317401897 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.413964    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.414018    4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.414033    4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:44:16 crc kubenswrapper[4824]: E0122 09:44:16.414116    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.414091539 +0000 UTC m=+153.317568602 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered]
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.455427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.455462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.455471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.455484    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.455493    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.557278    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.557339    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.557378    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.557408    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.557429    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.603379    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/2.log"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.607065    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.607596    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.627779    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.640104    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.653617    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.659662    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.659712    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.659724    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.659745    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.659756    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.671059    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.682238    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.693530    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.705748    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.718611    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.733814    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.744579    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.757476    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.761638    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.761672    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.761680    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.761696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.761706    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.768082    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.778316    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.790628    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.803023    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.816289    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.828286    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.841392    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:16Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.864424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.864479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.864491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.864508    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.864521    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.967581    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.967628    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.967640    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.967658    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:16 crc kubenswrapper[4824]: I0122 09:44:16.967671    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:16Z","lastTransitionTime":"2026-01-22T09:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.070476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.070530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.070540    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.070560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.070571    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.173224    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.173325    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.173349    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.173410    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.173433    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.276473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.276548    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.276565    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.276588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.276602    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.379381    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.379443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.379456    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.379479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.379492    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.410668    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 23:54:06.48108797 +0000 UTC
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.424673    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"]
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.425786    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.444239    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.460244    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.475638    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.482430    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.482515    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.482537    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.482559    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.482574    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.488292    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.505172    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.522175    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.537955    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.561161    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.579961    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.587440    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.587788    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.587905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.587993    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.588073    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.597677    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.615409    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/3.log"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.616180    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.616670    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/2.log"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.624316    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646" exitCode=1
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.624691    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.624828    4824 scope.go:117] "RemoveContainer" containerID="3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.625748    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:44:17 crc kubenswrapper[4824]: E0122 09:44:17.626049    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.641958    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.656938    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.676740    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.691558    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.691603    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.691618    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.691640    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.691654    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.698234    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.724126    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.739039    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.752554    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.768866    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.787184    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.794035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.794085    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.794098    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.794123    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.794135    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.804239    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.817982    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.832977    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.848340    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.870046    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3471ceac4d7220b2b8d30eb0aee2115db1c37d32d779663e5f3eaf02b6ac2373\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:43:44Z\\\",\\\"message\\\":\\\"cal_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0122 09:43:43.869644    6442 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0122 09:43:43.869674    6442 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default:  []services.lbConfig(nil)\\\\nI0122 09:43:43.869695    6442 services_controller.go:445] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF0122 09:43:43.869706    6442 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:17Z\\\",\\\"message\\\":\\\"nsole/networking-console-plugin-85b44fc459-gdk6g openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-4527v openshift-multus/multus-additional-cni-plugins-9wczc openshift-multus/network-metrics-daemon-xxtjz openshift-network-node-identity/network-node-identity-vrzqb openshift-image-registry/node-ca-tqrnv openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-zvvt5 openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k openshift-ovn-kubernetes/ovnkube-node-s8vt4]\\\\nI0122 09:44:16.840106    6913 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-image-registry/image-registry-operator. OVN-Kubernetes controller took 0.105489793 seconds. No OVN measurement.\\\\nI0122 09:44:16.839580    6913 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nF0122 09:44:16.840126    6913 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:44:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.886468    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.897993    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.898054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.898067    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.898090    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.898104    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:17Z","lastTransitionTime":"2026-01-22T09:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.901297    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.918116    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.936160    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.951629    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.965239    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.977096    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:17 crc kubenswrapper[4824]: I0122 09:44:17.988569    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:17Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.001278    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.001336    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.001365    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.001392    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.001409    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.007849    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47f8f412-e18a-4789-9553-eff66d7b53e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce73eb24a4e7c75b937de8520c772a2b37aa4e1f4aa142438701a243d1d2e3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af36f712c0370669b5316d98b277194b465ae3bd9197482859245a049ad0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a0eec4fb2313dfd42a94f3a70cc7a509038dd5e624597cb05961d1dc1e3849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://044fa372f704e0ec5d127058e2c208fbb15d0c7d1cba8497a9d495573e9ecdbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa91323c49714652c26af8de91efa2ccbc2025ff4b7197e48bfa4aed0792a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.021981    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.035383    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.103755    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.103809    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.103845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.103864    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.103876    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.206348    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.206434    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.206443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.206463    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.206474    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.309669    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.309723    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.309735    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.309756    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.309766    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.403498    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.403554    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.403593    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.403620    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:18 crc kubenswrapper[4824]: E0122 09:44:18.404401    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:18 crc kubenswrapper[4824]: E0122 09:44:18.404436    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:18 crc kubenswrapper[4824]: E0122 09:44:18.404653    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:18 crc kubenswrapper[4824]: E0122 09:44:18.404517    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.411797    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 21:55:07.537890096 +0000 UTC
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.413330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.413493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.413573    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.413659    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.413734    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.517330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.517412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.517427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.517447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.517461    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.620493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.620530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.620541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.620558    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.620570    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.629192    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/3.log"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.633997    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:44:18 crc kubenswrapper[4824]: E0122 09:44:18.634237    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.649841    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.665085    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.680496    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.699302    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.713820    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.724131    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.724189    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.724209    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.724231    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.724246    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.726275    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.740590    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.770202    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:17Z\\\",\\\"message\\\":\\\"nsole/networking-console-plugin-85b44fc459-gdk6g openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-4527v openshift-multus/multus-additional-cni-plugins-9wczc openshift-multus/network-metrics-daemon-xxtjz openshift-network-node-identity/network-node-identity-vrzqb openshift-image-registry/node-ca-tqrnv openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-zvvt5 openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k openshift-ovn-kubernetes/ovnkube-node-s8vt4]\\\\nI0122 09:44:16.840106    6913 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-image-registry/image-registry-operator. OVN-Kubernetes controller took 0.105489793 seconds. No OVN measurement.\\\\nI0122 09:44:16.839580    6913 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nF0122 09:44:16.840126    6913 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:44:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.783464    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.798748    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.810123    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.822277    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.826988    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.827046    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.827058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.827077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.827091    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.833107    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.844084    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.854607    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.866602    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.884532    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47f8f412-e18a-4789-9553-eff66d7b53e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce73eb24a4e7c75b937de8520c772a2b37aa4e1f4aa142438701a243d1d2e3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af36f712c0370669b5316d98b277194b465ae3bd9197482859245a049ad0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a0eec4fb2313dfd42a94f3a70cc7a509038dd5e624597cb05961d1dc1e3849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://044fa372f704e0ec5d127058e2c208fbb15d0c7d1cba8497a9d495573e9ecdbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa91323c49714652c26af8de91efa2ccbc2025ff4b7197e48bfa4aed0792a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.900665    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.913042    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:18Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.930499    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.930795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.930813    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.930838    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:18 crc kubenswrapper[4824]: I0122 09:44:18.930856    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:18Z","lastTransitionTime":"2026-01-22T09:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.033852    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.033895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.033905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.033921    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.033932    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.136960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.137039    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.137056    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.137082    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.137100    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.239509    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.239561    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.239611    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.239633    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.239643    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.342291    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.342326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.342334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.342348    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.342395    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.412270    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 22:50:01.723521273 +0000 UTC
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.445758    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.445841    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.445905    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.445932    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.445950    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.548177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.548214    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.548222    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.548237    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.548247    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.650444    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.650507    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.650524    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.650551    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.650570    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.753511    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.753562    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.753574    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.753595    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.753611    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.856563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.856602    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.856614    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.856632    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.856645    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.959274    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.959326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.959337    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.959411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:19 crc kubenswrapper[4824]: I0122 09:44:19.959424    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:19Z","lastTransitionTime":"2026-01-22T09:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.062869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.062913    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.062928    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.062946    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.062958    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.166009    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.166056    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.166067    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.166084    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.166097    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.269165    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.269208    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.269219    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.269235    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.269248    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.371801    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.371851    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.371863    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.371881    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.371894    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.403651    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.403651    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:20 crc kubenswrapper[4824]: E0122 09:44:20.403850    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.403682    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.403661    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:20 crc kubenswrapper[4824]: E0122 09:44:20.403992    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:20 crc kubenswrapper[4824]: E0122 09:44:20.404029    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:20 crc kubenswrapper[4824]: E0122 09:44:20.404216    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.412861    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 21:12:58.011556895 +0000 UTC
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.474628    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.474678    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.474694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.474719    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.474736    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.578090    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.578131    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.578143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.578165    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.578177    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.681116    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.681180    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.681207    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.681241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.681264    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.784393    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.784439    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.784453    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.784474    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.784489    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.887296    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.887383    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.887403    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.887429    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.887447    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.990333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.990434    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.990459    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.990491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:20 crc kubenswrapper[4824]: I0122 09:44:20.990513    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:20Z","lastTransitionTime":"2026-01-22T09:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.093259    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.093305    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.093316    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.093335    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.093379    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.196581    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.196638    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.196654    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.196678    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.196693    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.301690    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.301764    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.301782    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.301809    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.301826    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.405261    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.405325    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.405343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.405407    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.405433    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.413810    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 03:31:36.394107517 +0000 UTC
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.509536    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.509987    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.510164    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.510435    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.510647    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.614663    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.614773    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.614792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.614820    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.614838    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.717432    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.717505    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.717519    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.717540    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.717552    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.821180    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.821248    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.821264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.821290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.821307    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.923799    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.923858    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.923880    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.923909    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:21 crc kubenswrapper[4824]: I0122 09:44:21.923931    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:21Z","lastTransitionTime":"2026-01-22T09:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.026747    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.026828    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.026852    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.026886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.026908    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.129760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.129820    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.129835    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.129858    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.129875    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.232677    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.232745    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.232762    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.232787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.232804    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.336128    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.336173    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.336182    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.336200    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.336210    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.403559    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:22 crc kubenswrapper[4824]: E0122 09:44:22.403908    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.404117    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:22 crc kubenswrapper[4824]: E0122 09:44:22.404165    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.404265    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:22 crc kubenswrapper[4824]: E0122 09:44:22.404308    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.404450    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:22 crc kubenswrapper[4824]: E0122 09:44:22.404506    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.414604    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 10:59:56.632592798 +0000 UTC
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.439156    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.439222    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.439240    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.439264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.439331    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.542032    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.542069    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.542077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.542092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.542102    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.645671    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.645740    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.645762    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.645815    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.645828    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.748972    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.749041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.749055    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.749077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.749090    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.851713    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.851774    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.851790    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.851815    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.851834    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.955851    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.955921    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.955938    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.955964    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:22 crc kubenswrapper[4824]: I0122 09:44:22.955981    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:22Z","lastTransitionTime":"2026-01-22T09:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.060228    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.060305    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.060322    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.060348    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.060403    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.163778    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.163862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.163895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.163925    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.163946    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.267224    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.267292    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.267325    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.267400    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.267442    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.370700    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.370743    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.370755    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.370777    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.370788    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.415062    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 13:20:50.346653337 +0000 UTC
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.474492    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.474585    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.474608    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.474637    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.474658    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.577331    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.577450    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.577472    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.577494    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.577510    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.680176    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.680742    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.680760    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.680785    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.680804    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.732792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.732852    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.732874    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.732899    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.732921    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: E0122 09:44:23.745603    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:23Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.749446    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.749503    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.749528    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.749558    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.749581    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: E0122 09:44:23.761522    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:23Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.765345    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.765404    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.765412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.765427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.765439    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: E0122 09:44:23.775877    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:23Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.779547    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.779575    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.779583    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.779596    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.779604    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: E0122 09:44:23.791149    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:23Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.794283    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.794337    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.794385    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.794412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.794428    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: E0122 09:44:23.808229    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:23Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:23 crc kubenswrapper[4824]: E0122 09:44:23.808340    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.810631    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.810656    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.810665    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.810680    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.810693    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.913607    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.913671    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.913687    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.913712    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:23 crc kubenswrapper[4824]: I0122 09:44:23.913731    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:23Z","lastTransitionTime":"2026-01-22T09:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.016588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.016656    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.016677    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.016701    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.016718    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.119318    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.119387    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.119400    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.119420    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.119431    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.221577    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.221629    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.221647    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.221667    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.221678    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.325289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.325344    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.325391    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.325417    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.325437    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.403552    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:24 crc kubenswrapper[4824]: E0122 09:44:24.403723    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.404010    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:24 crc kubenswrapper[4824]: E0122 09:44:24.404112    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.404302    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:24 crc kubenswrapper[4824]: E0122 09:44:24.404433    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.404589    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:24 crc kubenswrapper[4824]: E0122 09:44:24.404733    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.415924    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 19:17:53.014499592 +0000 UTC
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.428260    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.428324    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.428350    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.428413    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.428436    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.532794    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.532843    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.532862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.532887    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.532904    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.636126    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.636243    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.636266    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.636296    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.636317    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.739422    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.739512    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.739619    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.739649    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.739664    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.842910    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.842992    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.843016    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.843048    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.843069    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.946278    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.946323    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.946335    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.946374    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:24 crc kubenswrapper[4824]: I0122 09:44:24.946384    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:24Z","lastTransitionTime":"2026-01-22T09:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.049788    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.049836    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.049847    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.049865    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.049877    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.153245    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.153302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.153319    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.153345    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.153385    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.256411    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.256464    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.256478    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.256496    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.256508    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.359532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.359618    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.359648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.359678    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.359701    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.416965    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 16:09:32.190202066 +0000 UTC
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.463054    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.463108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.463119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.463137    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.463149    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.566763    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.566799    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.566808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.566822    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.566831    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.669184    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.669239    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.669255    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.669278    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.669294    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.772388    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.772443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.772452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.772471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.772486    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.874627    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.874690    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.874707    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.874733    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.874752    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.977630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.977691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.977702    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.977720    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:25 crc kubenswrapper[4824]: I0122 09:44:25.977733    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:25Z","lastTransitionTime":"2026-01-22T09:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.080890    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.080938    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.080953    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.080969    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.080982    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.183277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.183323    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.183334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.183374    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.183392    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.286011    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.286105    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.286121    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.286143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.286160    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.389008    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.389064    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.389081    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.389104    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.389120    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.403917    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.403944    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.404023    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:26 crc kubenswrapper[4824]: E0122 09:44:26.404079    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.403927    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:26 crc kubenswrapper[4824]: E0122 09:44:26.404181    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:26 crc kubenswrapper[4824]: E0122 09:44:26.404283    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:26 crc kubenswrapper[4824]: E0122 09:44:26.404448    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.418196    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 04:48:35.170595729 +0000 UTC
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.491644    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.491692    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.491705    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.491726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.491741    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.596339    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.596450    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.596471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.596497    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.596515    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.699084    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.699131    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.699139    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.699156    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.699172    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.802475    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.802571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.802594    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.802626    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.802651    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.906267    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.906337    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.906346    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.906382    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:26 crc kubenswrapper[4824]: I0122 09:44:26.906393    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:26Z","lastTransitionTime":"2026-01-22T09:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.008987    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.009030    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.009042    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.009060    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.009073    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.111850    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.111932    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.111955    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.111983    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.111999    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.214793    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.214854    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.214870    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.214897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.214915    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.318253    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.318291    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.318299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.318313    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.318324    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.419017    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 05:22:25.491661937 +0000 UTC
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.419613    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.422121    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.422231    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.422277    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.422316    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.422410    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.444311    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.465644    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.485404    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.510588    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:17Z\\\",\\\"message\\\":\\\"nsole/networking-console-plugin-85b44fc459-gdk6g openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-4527v openshift-multus/multus-additional-cni-plugins-9wczc openshift-multus/network-metrics-daemon-xxtjz openshift-network-node-identity/network-node-identity-vrzqb openshift-image-registry/node-ca-tqrnv openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-zvvt5 openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k openshift-ovn-kubernetes/ovnkube-node-s8vt4]\\\\nI0122 09:44:16.840106    6913 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-image-registry/image-registry-operator. OVN-Kubernetes controller took 0.105489793 seconds. No OVN measurement.\\\\nI0122 09:44:16.839580    6913 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nF0122 09:44:16.840126    6913 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:44:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.524820    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.526139    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.526172    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.526183    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.526199    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.526429    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.535445    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.545977    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.570870    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47f8f412-e18a-4789-9553-eff66d7b53e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce73eb24a4e7c75b937de8520c772a2b37aa4e1f4aa142438701a243d1d2e3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af36f712c0370669b5316d98b277194b465ae3bd9197482859245a049ad0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a0eec4fb2313dfd42a94f3a70cc7a509038dd5e624597cb05961d1dc1e3849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://044fa372f704e0ec5d127058e2c208fbb15d0c7d1cba8497a9d495573e9ecdbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa91323c49714652c26af8de91efa2ccbc2025ff4b7197e48bfa4aed0792a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.602408    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.624181    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.628660    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.628811    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.628901    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.628968    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.629035    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.640766    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.651714    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.670380    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.683805    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.696436    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.707647    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.718402    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.730185    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:27Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.731486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.731518    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.731532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.731549    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.731561    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.834935    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.835270    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.835445    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.835806    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.835967    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.938190    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.938227    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.938240    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.938258    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:27 crc kubenswrapper[4824]: I0122 09:44:27.938269    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:27Z","lastTransitionTime":"2026-01-22T09:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.041113    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.041153    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.041165    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.041182    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.041194    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.144222    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.144266    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.144282    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.144302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.144318    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.247059    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.247334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.247483    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.247596    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.247696    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.350279    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.350332    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.350347    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.350391    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.350406    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.403422    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.403487    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:28 crc kubenswrapper[4824]: E0122 09:44:28.403630    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.403696    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:28 crc kubenswrapper[4824]: E0122 09:44:28.403803    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:28 crc kubenswrapper[4824]: E0122 09:44:28.404037    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.404298    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:28 crc kubenswrapper[4824]: E0122 09:44:28.404600    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.420010    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 05:04:57.620063953 +0000 UTC
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.458521    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.458581    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.458610    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.458635    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.458654    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.562264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.562315    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.562328    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.562346    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.562373    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.666018    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.666088    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.666111    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.666142    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.666164    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.768198    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.768239    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.768248    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.768265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.768275    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.870958    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.871001    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.871015    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.871032    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.871044    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.974117    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.974191    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.974208    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.974239    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:28 crc kubenswrapper[4824]: I0122 09:44:28.974257    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:28Z","lastTransitionTime":"2026-01-22T09:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.077457    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.077530    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.077555    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.077585    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.077606    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.180798    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.180851    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.180860    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.180877    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.180888    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.284119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.284167    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.284183    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.284207    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.284224    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.386630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.386694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.386713    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.386736    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.386753    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.420654    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 00:07:01.009811943 +0000 UTC
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.489217    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.489269    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.489279    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.489297    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.489309    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.592112    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.592184    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.592198    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.592219    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.592230    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.694752    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.694798    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.694811    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.694830    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.694843    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.797948    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.798045    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.798077    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.798110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.798134    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.901238    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.901286    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.901299    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.901316    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:29 crc kubenswrapper[4824]: I0122 09:44:29.901329    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:29Z","lastTransitionTime":"2026-01-22T09:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.004636    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.004696    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.004713    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.004737    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.004754    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.108114    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.108158    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.108170    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.108190    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.108204    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.211035    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.211089    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.211100    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.211123    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.211136    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.313880    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.313939    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.313955    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.313981    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.313997    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.403715    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:30 crc kubenswrapper[4824]: E0122 09:44:30.403920    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.404203    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:30 crc kubenswrapper[4824]: E0122 09:44:30.404291    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.404535    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:30 crc kubenswrapper[4824]: E0122 09:44:30.404631    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.404712    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:30 crc kubenswrapper[4824]: E0122 09:44:30.404928    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.417495    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.417556    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.417568    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.417589    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.417602    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.421643    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 04:37:22.981816015 +0000 UTC
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.520897    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.520943    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.520952    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.520969    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.520980    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.624212    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.624281    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.624300    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.624327    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.624344    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.727928    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.727977    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.727988    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.728006    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.728021    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.831493    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.831546    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.831568    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.831588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.831603    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.875658    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:30 crc kubenswrapper[4824]: E0122 09:44:30.875804    4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:44:30 crc kubenswrapper[4824]: E0122 09:44:30.875866    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs podName:f43b4618-291a-40d3-b1ca-32abfadaf376 nodeName:}" failed. No retries permitted until 2026-01-22 09:45:34.875846586 +0000 UTC m=+167.779323579 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs") pod "network-metrics-daemon-xxtjz" (UID: "f43b4618-291a-40d3-b1ca-32abfadaf376") : object "openshift-multus"/"metrics-daemon-secret" not registered
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.934625    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.934687    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.934705    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.934730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:30 crc kubenswrapper[4824]: I0122 09:44:30.934749    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:30Z","lastTransitionTime":"2026-01-22T09:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.037401    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.037473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.037490    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.037517    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.037535    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.140070    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.140108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.140118    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.140201    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.140217    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.243154    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.243209    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.243225    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.243246    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.243257    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.346380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.346422    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.346570    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.346590    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.346602    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.404575    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:44:31 crc kubenswrapper[4824]: E0122 09:44:31.404785    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.422846    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 08:07:33.718672555 +0000 UTC
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.450605    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.450709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.450726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.450751    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.450767    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.554088    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.554168    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.554180    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.554197    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.554213    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.658147    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.658218    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.658236    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.658265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.658283    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.760742    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.760802    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.760824    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.760847    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.760862    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.863291    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.863389    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.863404    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.863424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.863435    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.965371    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.965419    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.965429    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.965447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:31 crc kubenswrapper[4824]: I0122 09:44:31.965456    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:31Z","lastTransitionTime":"2026-01-22T09:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.068069    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.068114    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.068125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.068141    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.068151    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.171155    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.171194    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.171204    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.171222    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.171235    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.273862    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.273926    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.273940    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.273960    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.273972    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.376869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.376915    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.376927    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.376944    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.376955    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.403659    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.403781    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.403979    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.404028    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:32 crc kubenswrapper[4824]: E0122 09:44:32.404123    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:32 crc kubenswrapper[4824]: E0122 09:44:32.404213    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:32 crc kubenswrapper[4824]: E0122 09:44:32.404323    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:32 crc kubenswrapper[4824]: E0122 09:44:32.404528    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.423707    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 15:21:47.342226346 +0000 UTC
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.480288    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.480324    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.480334    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.480365    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.480379    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.582787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.582836    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.582852    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.582875    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.582890    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.686125    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.686193    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.686204    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.686223    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.686239    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.789923    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.789984    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.789997    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.790021    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.790035    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.893226    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.893264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.893274    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.893289    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.893298    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.997293    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.997424    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.997451    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.997496    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:32 crc kubenswrapper[4824]: I0122 09:44:32.997524    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:32Z","lastTransitionTime":"2026-01-22T09:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.100157    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.100218    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.100233    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.100257    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.100273    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.203036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.203098    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.203115    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.203144    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.203163    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.313303    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.313593    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.314160    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.314324    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.314381    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.417062    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.417134    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.417151    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.417177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.417195    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.424284    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 01:38:01.780829379 +0000 UTC
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.520388    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.520433    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.520452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.520471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.520484    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.623196    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.623248    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.623265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.623286    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.623301    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.727012    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.727090    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.727108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.727137    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.727156    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.830817    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.830880    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.830895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.830916    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.830931    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.934510    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.934588    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.934607    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.934633    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:33 crc kubenswrapper[4824]: I0122 09:44:33.934650    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:33Z","lastTransitionTime":"2026-01-22T09:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.037694    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.037754    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.037776    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.037805    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.037825    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.140430    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.140505    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.140542    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.140572    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.140628    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.142571    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.142630    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.142647    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.142667    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.142682    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.164331    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:34Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.168947    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.168991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.169012    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.169042    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.169064    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.186052    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:34Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.190709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.190756    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.190778    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.190808    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.190827    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.208895    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:34Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.214327    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.214418    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.214437    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.214462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.214482    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.232717    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:34Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.237991    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.238051    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.238069    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.238094    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.238113    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.254838    4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b28d29c1-bb01-432c-b92a-540ec89f6524\\\",\\\"systemUUID\\\":\\\"2a2d1f9c-78d4-4a51-9bdc-24b710991b2b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:34Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.255158    4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.257576    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.257691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.257778    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.257816    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.257838    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.360429    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.360488    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.360509    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.360541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.360564    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.403345    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.403512    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.403577    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.403596    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.403573    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.403823    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.404385    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:34 crc kubenswrapper[4824]: E0122 09:44:34.404452    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.425248    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:23:24.026543011 +0000 UTC
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.463726    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.463792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.463806    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.463827    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.463845    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.567105    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.567177    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.567192    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.567215    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.567231    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.676245    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.676324    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.676342    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.676400    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.676426    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.779410    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.779462    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.779471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.779491    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.779502    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.882946    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.883010    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.883033    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.883062    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.883083    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.985269    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.985326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.985349    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.985422    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:34 crc kubenswrapper[4824]: I0122 09:44:34.985450    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:34Z","lastTransitionTime":"2026-01-22T09:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.088182    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.088218    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.088232    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.088264    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.088275    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.191052    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.191124    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.191146    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.191176    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.191200    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.294680    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.294717    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.294728    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.294745    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.294755    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.396563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.396618    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.396636    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.396659    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.396675    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.425760    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 03:52:09.616696718 +0000 UTC
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.499683    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.499789    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.499845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.499869    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.499885    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.602470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.602513    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.602522    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.602539    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.602549    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.705573    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.705658    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.705691    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.705723    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.705748    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.808814    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.808882    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.808895    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.808916    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.808928    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.912246    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.912414    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.912441    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.912473    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:35 crc kubenswrapper[4824]: I0122 09:44:35.912495    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:35Z","lastTransitionTime":"2026-01-22T09:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.015697    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.015771    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.015793    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.015822    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.015843    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.119171    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.119207    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.119215    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.119232    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.119281    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.222522    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.222597    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.222611    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.222635    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.222652    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.324638    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.324701    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.324713    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.324734    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.324746    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.404004    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.404000    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.404026    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.404209    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:36 crc kubenswrapper[4824]: E0122 09:44:36.404298    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:36 crc kubenswrapper[4824]: E0122 09:44:36.404644    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:36 crc kubenswrapper[4824]: E0122 09:44:36.404947    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:36 crc kubenswrapper[4824]: E0122 09:44:36.405058    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.425922    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 05:57:54.004105868 +0000 UTC
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.427692    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.427792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.427813    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.427841    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.427859    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.531168    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.531230    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.531248    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.531270    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.531285    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.674730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.674781    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.674793    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.674810    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.674821    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.777961    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.778221    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.778269    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.778292    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.778311    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.880951    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.881082    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.881113    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.881146    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.881169    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.984333    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.984447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.984459    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.984476    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:36 crc kubenswrapper[4824]: I0122 09:44:36.984491    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:36Z","lastTransitionTime":"2026-01-22T09:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.087693    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.087772    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.087792    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.087821    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.087838    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.191198    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.191246    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.191255    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.191271    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.191284    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.294010    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.294100    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.294126    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.294166    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.294190    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.397161    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.397218    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.397233    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.397254    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.397267    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.426081    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 01:12:02.141581273 +0000 UTC
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.431921    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9wczc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68893d7c-efac-44f9-8935-843169e413ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8155c405da745f900ea2d8b4907faaa13db3ac48b9d23510e05b35eb4c2041d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dfe1db8ae18402b752e883fee6f8145ec4d394ea7993f4b5a0de5341cff6e55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1455662ac38d64b12c7c04a939c1b5f1ac1921a1f028fd9ea4d7d1c1d1e228bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877bd996dd33cfcf1dba58c788baf06075e55b9690dcb2881fd6087b363690aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5087bfa675b1db1845dea142b889a922728aa11f137238da243a4c0130dfd91c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc96591bd82109394dd53f252fa0414996798ecf5832c5155c3dc1b2a7eee87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cb04333b73f71b28c3e6757abac00ec47059b4573bbc029d47a4f1d41f7f006\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qj5rl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9wczc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.466640    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:17Z\\\",\\\"message\\\":\\\"nsole/networking-console-plugin-85b44fc459-gdk6g openshift-network-operator/network-operator-58b4c7f79c-55gtf openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/machine-config-daemon-4527v openshift-multus/multus-additional-cni-plugins-9wczc openshift-multus/network-metrics-daemon-xxtjz openshift-network-node-identity/network-node-identity-vrzqb openshift-image-registry/node-ca-tqrnv openshift-kube-controller-manager/kube-controller-manager-crc openshift-multus/multus-zvvt5 openshift-network-diagnostics/network-check-source-55646444c4-trplf openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k openshift-ovn-kubernetes/ovnkube-node-s8vt4]\\\\nI0122 09:44:16.840106    6913 ovnkube_controller.go:1292] Config duration recorder: kind/namespace/name service/openshift-image-registry/image-registry-operator. OVN-Kubernetes controller took 0.105489793 seconds. No OVN measurement.\\\\nI0122 09:44:16.839580    6913 services_controller.go:356] Processing sync for service openshift-authentication/oauth-openshift for network=default\\\\nF0122 09:44:16.840126    6913 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller ini\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:44:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5pq4s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:13Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s8vt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.482262    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f43b4618-291a-40d3-b1ca-32abfadaf376\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jjtgq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:26Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xxtjz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.499499    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.499541    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.499551    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.499569    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.499581    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.501742    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://781314ef9085823cc73f5ce19fb08d6026ad7a35efae740be8d48a4f1de4fb26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6fd6f9ad63378397c63ce4a07ed13fa150edb904822397b38a6ed895be769c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.516561    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.531950    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac403bb6c9fbc7e395b9764aacf24513a6229757c32900dc72804b03ca8f581e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.543576    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db09d5bd73faae771d2d561eeda59bd1799a97428ad2caa4c11d197a8eed8aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.556868    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mdbj5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"798e70c0-4438-46c9-9efc-be9121d1e60e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49e877a409b159d281a195a753b5914407af438b5422f96669d25559bc51f717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zwdch\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mdbj5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.570391    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e22c2c99-5647-4fc8-9550-7e2a82584d55\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb6b66fba318e77fbe7d3e4065845f266fbed0bcdd80781f41b9973fcb34f86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqtpk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4527v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.588897    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"606a4ae4-6d9d-46e1-a1c3-583b4737af6b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e3de9e29c8e93e9efe9b5bbfe7137ea8a61d6a391c2bbb0e154fa56c8b6baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eebd92eae568beb0141dc2d8e02051cf2a316b5dc4866beb80b15012cc7232f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-brwsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-9948k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.603036    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.603065    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.603075    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.603092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.603102    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.612079    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47f8f412-e18a-4789-9553-eff66d7b53e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce73eb24a4e7c75b937de8520c772a2b37aa4e1f4aa142438701a243d1d2e3a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37af36f712c0370669b5316d98b277194b465ae3bd9197482859245a049ad0e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0a0eec4fb2313dfd42a94f3a70cc7a509038dd5e624597cb05961d1dc1e3849\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://044fa372f704e0ec5d127058e2c208fbb15d0c7d1cba8497a9d495573e9ecdbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa91323c49714652c26af8de91efa2ccbc2025ff4b7197e48bfa4aed0792a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d0db2d0fce0536f033a625a68b220ee086061a4438ca371663b07d568b1d9cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf041834484a246532f72e3b680e8f25660eec08f163f83f743b8514aa49dc7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b784a1ff38cb074910c89a73deb81ecebf2b946c020d76fac6b50b3b4d8b15f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.630982    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e5d7f0-3820-4ff5-8141-e66990987acf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0122 09:43:00.935206       1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0122 09:43:00.936792       1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4291149713/tls.crt::/tmp/serving-cert-4291149713/tls.key\\\\\\\"\\\\nI0122 09:43:11.570946       1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0122 09:43:11.575613       1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0122 09:43:11.575651       1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0122 09:43:11.575706       1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0122 09:43:11.575718       1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0122 09:43:11.585124       1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0122 09:43:11.585154       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0122 09:43:11.585155       1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0122 09:43:11.585159       1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0122 09:43:11.585177       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0122 09:43:11.585180       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0122 09:43:11.585184       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0122 09:43:11.585187       1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0122 09:43:11.588578       1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.642945    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8869792-684d-4678-bd9e-71432b5b26bc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://557d66554bdc1d32ddb3565c0d20be49d8d8cd0489b6f777b7b4741b78d6e7fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4706baf0838924935996a001695fbe155dd09ff5040f9f700cbf552aed377048\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.657152    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tqrnv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cced567c-003b-445f-b568-045b5a667853\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a6b4e8c1390d2abe3571b2aedd11c112ce0efbdfea31ba5322c6ffe347feb4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-896ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tqrnv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.670627    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.686541    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted.  The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.702598    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-zvvt5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-22T09:44:00Z\\\",\\\"message\\\":\\\"2026-01-22T09:43:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4\\\\n2026-01-22T09:43:14+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_606556e8-7dfd-4070-9ee4-995d438702d4 to /host/opt/cni/bin/\\\\n2026-01-22T09:43:14Z [verbose] multus-daemon started\\\\n2026-01-22T09:43:14Z [verbose] Readiness Indicator file check\\\\n2026-01-22T09:43:59Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-22T09:43:13Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t72pd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:43:12Z\\\"}}\" for pod \"openshift-multus\"/\"multus-zvvt5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.705019    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.705041    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.705050    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.705063    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.705073    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.716499    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d40cbe-a275-4dc6-bbb6-d8df363022e3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d77a95c0d0a630014e404cbdc955b9c0a10b05b46adbc16b812adbeef2ecc28a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69a73df559bc47dd5632446109d515fa95f301da66f47038c71b7980581f067c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0157483ff58fcccb66431ee297e4df47a6fbfc052e2350786bd423dd6b49b51\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.728796    4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c66913eb-1658-45f8-923e-f33e3e5db322\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-22T09:42:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a7b13c2000e308db2888bc13e510bfeb4b66ea1d4d2a35a0f0e8139ecf438a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3515d2cfdf891346866b7ec1ed2a9e25d5965b3c725896c64752f5260d6afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbd9c65061a04aebb8edf3f776448230b2675d322e937c77acacbabf7eb7919e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-22T09:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a587ac2f761abb55dbdc3480d48ed088bcd3a29d6e6049cd6ea83c0ecb51544f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-22T09:42:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-22T09:42:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-22T09:42:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-22T09:44:37Z is after 2025-08-24T17:21:41Z"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.807945    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.808032    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.808053    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.808071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.808083    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.913241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.913311    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.913331    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.913380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:37 crc kubenswrapper[4824]: I0122 09:44:37.913399    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:37Z","lastTransitionTime":"2026-01-22T09:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.015747    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.015795    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.015803    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.015822    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.015834    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.118714    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.118759    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.118768    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.118787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.118798    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.221603    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.221665    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.221684    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.221708    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.221724    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.325301    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.325394    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.325412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.325439    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.325459    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.403665    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.403723    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.403693    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.403694    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:38 crc kubenswrapper[4824]: E0122 09:44:38.403900    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:38 crc kubenswrapper[4824]: E0122 09:44:38.404023    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:38 crc kubenswrapper[4824]: E0122 09:44:38.404200    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:38 crc kubenswrapper[4824]: E0122 09:44:38.404335    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.426251    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 10:28:50.328585909 +0000 UTC
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.428455    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.428521    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.428544    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.428572    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.428594    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.531560    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.531667    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.531701    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.531730    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.531750    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.634648    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.634722    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.634738    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.634766    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.634785    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.738412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.738483    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.738509    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.738543    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.738568    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.842311    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.842410    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.842429    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.842454    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.842477    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.946019    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.946078    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.946093    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.946119    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:38 crc kubenswrapper[4824]: I0122 09:44:38.946135    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:38Z","lastTransitionTime":"2026-01-22T09:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.049443    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.049532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.049552    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.049577    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.049598    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.152667    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.152728    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.152744    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.152764    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.152776    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.256167    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.256225    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.256237    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.256298    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.256318    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.359528    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.359613    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.359633    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.359661    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.359680    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.427427    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 11:28:19.045290807 +0000 UTC
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.463294    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.463407    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.463427    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.463458    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.463477    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.566498    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.566586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.566599    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.566625    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.566639    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.669321    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.669425    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.669441    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.669468    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.669485    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.772531    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.772589    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.772605    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.772641    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.772673    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.876198    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.876268    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.876308    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.876344    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.876441    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.979549    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.979622    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.979641    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.979668    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:39 crc kubenswrapper[4824]: I0122 09:44:39.979687    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:39Z","lastTransitionTime":"2026-01-22T09:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.083470    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.083563    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.083585    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.083615    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.083635    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.186108    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.186173    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.186191    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.186217    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.186236    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.289452    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.289532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.289557    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.289590    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.289614    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.393326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.393486    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.393504    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.393531    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.393555    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.403806    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.403870    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.403838    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:40 crc kubenswrapper[4824]: E0122 09:44:40.404022    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.403810    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:40 crc kubenswrapper[4824]: E0122 09:44:40.404143    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:40 crc kubenswrapper[4824]: E0122 09:44:40.404334    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:40 crc kubenswrapper[4824]: E0122 09:44:40.404469    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.427660    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 09:11:49.202665178 +0000 UTC
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.496663    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.496725    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.496742    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.496766    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.496783    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.600447    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.600532    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.600550    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.600579    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.600603    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.704652    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.704734    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.704762    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.704791    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.704813    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.808180    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.808241    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.808255    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.808283    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.808300    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.911058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.911132    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.911143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.911176    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:40 crc kubenswrapper[4824]: I0122 09:44:40.911192    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:40Z","lastTransitionTime":"2026-01-22T09:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.014316    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.014407    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.014420    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.014438    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.014451    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.118489    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.118557    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.118575    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.118601    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.118620    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.222028    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.222098    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.222110    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.222129    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.222141    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.326690    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.326757    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.326772    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.326796    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.326810    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.427822    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 10:51:25.324578464 +0000 UTC
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.429737    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.429790    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.429805    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.429829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.429846    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.532717    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.532773    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.532787    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.532807    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.532819    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.635302    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.635343    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.635398    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.635417    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.635429    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.737193    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.737240    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.737255    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.737275    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.737287    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.840128    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.840203    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.840228    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.840262    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.840291    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.943829    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.943893    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.943912    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.943936    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:41 crc kubenswrapper[4824]: I0122 09:44:41.943951    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:41Z","lastTransitionTime":"2026-01-22T09:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.047487    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.047794    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.047826    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.047883    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.047910    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.151185    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.151266    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.151283    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.151304    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.151320    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.254256    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.254315    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.254330    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.254390    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.254408    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.358024    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.358106    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.358133    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.358167    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.358192    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.403602    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.403739    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:42 crc kubenswrapper[4824]: E0122 09:44:42.403801    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.403619    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.403838    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:42 crc kubenswrapper[4824]: E0122 09:44:42.404233    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:42 crc kubenswrapper[4824]: E0122 09:44:42.404414    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:42 crc kubenswrapper[4824]: E0122 09:44:42.404568    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.404857    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:44:42 crc kubenswrapper[4824]: E0122 09:44:42.405104    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s8vt4_openshift-ovn-kubernetes(40c9681a-fa96-4b9f-8ff5-2d81bbf190fb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.428942    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 23:00:30.19831647 +0000 UTC
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.461616    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.461689    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.461709    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.461734    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.461754    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.564758    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.564811    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.564821    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.564842    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.564858    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.668323    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.668459    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.668490    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.668561    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.668585    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.772756    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.772824    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.772849    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.772882    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.772908    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.875906    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.876172    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.876483    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.876574    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.876679    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.978845    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.978904    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.978914    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.978929    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:42 crc kubenswrapper[4824]: I0122 09:44:42.978939    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:42Z","lastTransitionTime":"2026-01-22T09:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.082007    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.082057    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.082066    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.082083    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.082096    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.185003    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.185052    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.185071    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.185092    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.185105    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.288677    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.288771    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.288810    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.288853    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.288878    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.392199    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.392250    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.392300    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.392326    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.392341    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.429066    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 19:10:53.608567119 +0000 UTC
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.495886    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.495962    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.495982    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.496014    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.496038    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.599284    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.599340    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.599388    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.599412    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.599426    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.702227    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.702265    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.702273    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.702290    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.702300    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.805764    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.805818    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.805831    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.805850    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.805864    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.908983    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.909044    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.909058    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.909086    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:43 crc kubenswrapper[4824]: I0122 09:44:43.909102    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:43Z","lastTransitionTime":"2026-01-22T09:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.011863    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.012380    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.012471    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.012654    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.012719    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:44Z","lastTransitionTime":"2026-01-22T09:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.116189    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.116282    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.116309    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.116342    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.116411    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:44Z","lastTransitionTime":"2026-01-22T09:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.219023    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.219065    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.219073    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.219088    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.219098    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:44Z","lastTransitionTime":"2026-01-22T09:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.322479    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.322586    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.322603    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.322626    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.322640    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:44Z","lastTransitionTime":"2026-01-22T09:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.404146    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.404237    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.404278    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.404391    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:44 crc kubenswrapper[4824]: E0122 09:44:44.404510    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:44 crc kubenswrapper[4824]: E0122 09:44:44.404608    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:44 crc kubenswrapper[4824]: E0122 09:44:44.404702    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:44 crc kubenswrapper[4824]: E0122 09:44:44.404841    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.421093    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.421132    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.421143    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.421161    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.421174    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:44Z","lastTransitionTime":"2026-01-22T09:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.429963    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 03:11:48.698555962 +0000 UTC
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.450565    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.450620    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.450633    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.450651    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.450662    4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-22T09:44:44Z","lastTransitionTime":"2026-01-22T09:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"}
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.477715    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"]
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.478336    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.480843    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.481776    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.482996    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.485822    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.507777    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=93.507758799 podStartE2EDuration="1m33.507758799s" podCreationTimestamp="2026-01-22 09:43:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.506036046 +0000 UTC m=+117.409513039" watchObservedRunningTime="2026-01-22 09:44:44.507758799 +0000 UTC m=+117.411235792"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.520460    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=68.520421441 podStartE2EDuration="1m8.520421441s" podCreationTimestamp="2026-01-22 09:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.518859373 +0000 UTC m=+117.422336366" watchObservedRunningTime="2026-01-22 09:44:44.520421441 +0000 UTC m=+117.423898454"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.565671    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-zvvt5" podStartSLOduration=92.56564443 podStartE2EDuration="1m32.56564443s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.565192376 +0000 UTC m=+117.468669369" watchObservedRunningTime="2026-01-22 09:44:44.56564443 +0000 UTC m=+117.469121423"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.614041    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9wczc" podStartSLOduration=92.614013916 podStartE2EDuration="1m32.614013916s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.613206761 +0000 UTC m=+117.516683754" watchObservedRunningTime="2026-01-22 09:44:44.614013916 +0000 UTC m=+117.517490909"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.665649    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-9948k" podStartSLOduration=92.665623773 podStartE2EDuration="1m32.665623773s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.664760936 +0000 UTC m=+117.568237929" watchObservedRunningTime="2026-01-22 09:44:44.665623773 +0000 UTC m=+117.569100756"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.668344    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/51d37a90-320c-41a5-86e6-3906a17b0900-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.668414    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/51d37a90-320c-41a5-86e6-3906a17b0900-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.668445    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d37a90-320c-41a5-86e6-3906a17b0900-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.668469    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/51d37a90-320c-41a5-86e6-3906a17b0900-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.668493    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51d37a90-320c-41a5-86e6-3906a17b0900-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.717170    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=27.717149116999998 podStartE2EDuration="27.717149117s" podCreationTimestamp="2026-01-22 09:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.716851658 +0000 UTC m=+117.620328651" watchObservedRunningTime="2026-01-22 09:44:44.717149117 +0000 UTC m=+117.620626110"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.734492    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.734470813 podStartE2EDuration="1m32.734470813s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.732750109 +0000 UTC m=+117.636227112" watchObservedRunningTime="2026-01-22 09:44:44.734470813 +0000 UTC m=+117.637947806"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769076    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d37a90-320c-41a5-86e6-3906a17b0900-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769120    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/51d37a90-320c-41a5-86e6-3906a17b0900-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769149    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51d37a90-320c-41a5-86e6-3906a17b0900-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769171    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/51d37a90-320c-41a5-86e6-3906a17b0900-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769198    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/51d37a90-320c-41a5-86e6-3906a17b0900-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769251    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/51d37a90-320c-41a5-86e6-3906a17b0900-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.769452    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/51d37a90-320c-41a5-86e6-3906a17b0900-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.770565    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/51d37a90-320c-41a5-86e6-3906a17b0900-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.776405    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51d37a90-320c-41a5-86e6-3906a17b0900-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.786260    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51d37a90-320c-41a5-86e6-3906a17b0900-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s95kc\" (UID: \"51d37a90-320c-41a5-86e6-3906a17b0900\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.799881    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.800333    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podStartSLOduration=92.80031891 podStartE2EDuration="1m32.80031891s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.799594037 +0000 UTC m=+117.703071030" watchObservedRunningTime="2026-01-22 09:44:44.80031891 +0000 UTC m=+117.703795903"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.801087    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mdbj5" podStartSLOduration=92.801079323 podStartE2EDuration="1m32.801079323s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.786501842 +0000 UTC m=+117.689978845" watchObservedRunningTime="2026-01-22 09:44:44.801079323 +0000 UTC m=+117.704556316"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.812433    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=47.812406044 podStartE2EDuration="47.812406044s" podCreationTimestamp="2026-01-22 09:43:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.811585468 +0000 UTC m=+117.715062461" watchObservedRunningTime="2026-01-22 09:44:44.812406044 +0000 UTC m=+117.715883057"
Jan 22 09:44:44 crc kubenswrapper[4824]: I0122 09:44:44.827551    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tqrnv" podStartSLOduration=92.827528521 podStartE2EDuration="1m32.827528521s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:44.826939443 +0000 UTC m=+117.730416436" watchObservedRunningTime="2026-01-22 09:44:44.827528521 +0000 UTC m=+117.731005514"
Jan 22 09:44:45 crc kubenswrapper[4824]: I0122 09:44:45.431004    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 17:10:49.67068224 +0000 UTC
Jan 22 09:44:45 crc kubenswrapper[4824]: I0122 09:44:45.431103    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates
Jan 22 09:44:45 crc kubenswrapper[4824]: I0122 09:44:45.442148    4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146
Jan 22 09:44:45 crc kubenswrapper[4824]: I0122 09:44:45.732323    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc" event={"ID":"51d37a90-320c-41a5-86e6-3906a17b0900","Type":"ContainerStarted","Data":"1a7dd68ef2dcf8ae076924407c06a9c5145c268a4c36b0888721cfbca2297841"}
Jan 22 09:44:45 crc kubenswrapper[4824]: I0122 09:44:45.732859    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc" event={"ID":"51d37a90-320c-41a5-86e6-3906a17b0900","Type":"ContainerStarted","Data":"678b0b190f92b80c4426f6b2c0e4099d07ba95c5033125c6bcd08ada80673be0"}
Jan 22 09:44:45 crc kubenswrapper[4824]: I0122 09:44:45.751498    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s95kc" podStartSLOduration=93.751471692 podStartE2EDuration="1m33.751471692s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:45.750718778 +0000 UTC m=+118.654195771" watchObservedRunningTime="2026-01-22 09:44:45.751471692 +0000 UTC m=+118.654948675"
Jan 22 09:44:46 crc kubenswrapper[4824]: I0122 09:44:46.404070    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:46 crc kubenswrapper[4824]: I0122 09:44:46.404200    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:46 crc kubenswrapper[4824]: I0122 09:44:46.404282    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:46 crc kubenswrapper[4824]: I0122 09:44:46.404296    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:46 crc kubenswrapper[4824]: E0122 09:44:46.404217    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:46 crc kubenswrapper[4824]: E0122 09:44:46.404542    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:46 crc kubenswrapper[4824]: E0122 09:44:46.404633    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:46 crc kubenswrapper[4824]: E0122 09:44:46.404714    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:47 crc kubenswrapper[4824]: E0122 09:44:47.342369    4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup"
Jan 22 09:44:47 crc kubenswrapper[4824]: E0122 09:44:47.498840    4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"
Jan 22 09:44:47 crc kubenswrapper[4824]: I0122 09:44:47.741219    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/1.log"
Jan 22 09:44:47 crc kubenswrapper[4824]: I0122 09:44:47.741980    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/0.log"
Jan 22 09:44:47 crc kubenswrapper[4824]: I0122 09:44:47.742157    4824 generic.go:334] "Generic (PLEG): container finished" podID="00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c" containerID="814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914" exitCode=1
Jan 22 09:44:47 crc kubenswrapper[4824]: I0122 09:44:47.742285    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerDied","Data":"814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914"}
Jan 22 09:44:47 crc kubenswrapper[4824]: I0122 09:44:47.742432    4824 scope.go:117] "RemoveContainer" containerID="660ddabf96647cee4d8b605aa75090bc969dfe838ba79a01901fb87482a152df"
Jan 22 09:44:47 crc kubenswrapper[4824]: I0122 09:44:47.742870    4824 scope.go:117] "RemoveContainer" containerID="814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914"
Jan 22 09:44:47 crc kubenswrapper[4824]: E0122 09:44:47.743142    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-zvvt5_openshift-multus(00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c)\"" pod="openshift-multus/multus-zvvt5" podUID="00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c"
Jan 22 09:44:48 crc kubenswrapper[4824]: I0122 09:44:48.403637    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:48 crc kubenswrapper[4824]: I0122 09:44:48.403701    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:48 crc kubenswrapper[4824]: I0122 09:44:48.403732    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:48 crc kubenswrapper[4824]: I0122 09:44:48.403807    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:48 crc kubenswrapper[4824]: E0122 09:44:48.404283    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:48 crc kubenswrapper[4824]: E0122 09:44:48.404475    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:48 crc kubenswrapper[4824]: E0122 09:44:48.404576    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:48 crc kubenswrapper[4824]: E0122 09:44:48.404655    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:48 crc kubenswrapper[4824]: I0122 09:44:48.747615    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/1.log"
Jan 22 09:44:50 crc kubenswrapper[4824]: I0122 09:44:50.403929    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:50 crc kubenswrapper[4824]: E0122 09:44:50.404171    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:50 crc kubenswrapper[4824]: I0122 09:44:50.404951    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:50 crc kubenswrapper[4824]: I0122 09:44:50.405035    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:50 crc kubenswrapper[4824]: I0122 09:44:50.405053    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:50 crc kubenswrapper[4824]: E0122 09:44:50.405668    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:50 crc kubenswrapper[4824]: E0122 09:44:50.405790    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:50 crc kubenswrapper[4824]: E0122 09:44:50.405851    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:52 crc kubenswrapper[4824]: I0122 09:44:52.403430    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:52 crc kubenswrapper[4824]: I0122 09:44:52.403480    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:52 crc kubenswrapper[4824]: I0122 09:44:52.403510    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:52 crc kubenswrapper[4824]: I0122 09:44:52.403593    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:52 crc kubenswrapper[4824]: E0122 09:44:52.403605    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:52 crc kubenswrapper[4824]: E0122 09:44:52.403732    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:52 crc kubenswrapper[4824]: E0122 09:44:52.403885    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:52 crc kubenswrapper[4824]: E0122 09:44:52.404185    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:52 crc kubenswrapper[4824]: E0122 09:44:52.504733    4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"
Jan 22 09:44:54 crc kubenswrapper[4824]: I0122 09:44:54.404225    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:54 crc kubenswrapper[4824]: I0122 09:44:54.404280    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:54 crc kubenswrapper[4824]: I0122 09:44:54.404292    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:54 crc kubenswrapper[4824]: I0122 09:44:54.404308    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:54 crc kubenswrapper[4824]: E0122 09:44:54.404506    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:54 crc kubenswrapper[4824]: E0122 09:44:54.404619    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:54 crc kubenswrapper[4824]: E0122 09:44:54.404720    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:54 crc kubenswrapper[4824]: E0122 09:44:54.404789    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:56 crc kubenswrapper[4824]: I0122 09:44:56.403737    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:56 crc kubenswrapper[4824]: I0122 09:44:56.403973    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:56 crc kubenswrapper[4824]: I0122 09:44:56.404525    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:56 crc kubenswrapper[4824]: I0122 09:44:56.403973    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:56 crc kubenswrapper[4824]: E0122 09:44:56.404713    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:56 crc kubenswrapper[4824]: E0122 09:44:56.404959    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:56 crc kubenswrapper[4824]: E0122 09:44:56.405127    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:56 crc kubenswrapper[4824]: E0122 09:44:56.405285    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:57 crc kubenswrapper[4824]: I0122 09:44:57.407334    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:44:57 crc kubenswrapper[4824]: E0122 09:44:57.505745    4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.403101    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:44:58 crc kubenswrapper[4824]: E0122 09:44:58.403256    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.403500    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:44:58 crc kubenswrapper[4824]: E0122 09:44:58.403565    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.403683    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.404015    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:44:58 crc kubenswrapper[4824]: E0122 09:44:58.404075    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:44:58 crc kubenswrapper[4824]: E0122 09:44:58.404259    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.784848    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/3.log"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.787433    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerStarted","Data":"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"}
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.787899    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.813796    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podStartSLOduration=106.813768247 podStartE2EDuration="1m46.813768247s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:44:58.810018931 +0000 UTC m=+131.713495954" watchObservedRunningTime="2026-01-22 09:44:58.813768247 +0000 UTC m=+131.717245270"
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.944975    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xxtjz"]
Jan 22 09:44:58 crc kubenswrapper[4824]: I0122 09:44:58.945130    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:44:58 crc kubenswrapper[4824]: E0122 09:44:58.945253    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:45:00 crc kubenswrapper[4824]: I0122 09:45:00.403227    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:00 crc kubenswrapper[4824]: E0122 09:45:00.403439    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:45:00 crc kubenswrapper[4824]: I0122 09:45:00.403231    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:00 crc kubenswrapper[4824]: E0122 09:45:00.403518    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:45:00 crc kubenswrapper[4824]: I0122 09:45:00.403265    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:00 crc kubenswrapper[4824]: E0122 09:45:00.403570    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:45:00 crc kubenswrapper[4824]: I0122 09:45:00.403232    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:00 crc kubenswrapper[4824]: E0122 09:45:00.403614    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:45:02 crc kubenswrapper[4824]: I0122 09:45:02.403262    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:02 crc kubenswrapper[4824]: I0122 09:45:02.403395    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:02 crc kubenswrapper[4824]: I0122 09:45:02.403416    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:02 crc kubenswrapper[4824]: E0122 09:45:02.403481    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:45:02 crc kubenswrapper[4824]: I0122 09:45:02.403574    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:02 crc kubenswrapper[4824]: E0122 09:45:02.403687    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:45:02 crc kubenswrapper[4824]: E0122 09:45:02.404136    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:45:02 crc kubenswrapper[4824]: E0122 09:45:02.404388    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:45:02 crc kubenswrapper[4824]: I0122 09:45:02.404517    4824 scope.go:117] "RemoveContainer" containerID="814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914"
Jan 22 09:45:02 crc kubenswrapper[4824]: E0122 09:45:02.507931    4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"
Jan 22 09:45:03 crc kubenswrapper[4824]: I0122 09:45:03.806324    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/1.log"
Jan 22 09:45:03 crc kubenswrapper[4824]: I0122 09:45:03.806429    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerStarted","Data":"158628e706616c42e59547ac0e862beff096705bca967864fa84347bcd7b8aa5"}
Jan 22 09:45:04 crc kubenswrapper[4824]: I0122 09:45:04.404052    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:04 crc kubenswrapper[4824]: I0122 09:45:04.404158    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:04 crc kubenswrapper[4824]: I0122 09:45:04.404077    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:04 crc kubenswrapper[4824]: I0122 09:45:04.404240    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:04 crc kubenswrapper[4824]: E0122 09:45:04.404488    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:45:04 crc kubenswrapper[4824]: E0122 09:45:04.404597    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:45:04 crc kubenswrapper[4824]: E0122 09:45:04.404808    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:45:04 crc kubenswrapper[4824]: E0122 09:45:04.405111    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:45:06 crc kubenswrapper[4824]: I0122 09:45:06.403677    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:06 crc kubenswrapper[4824]: I0122 09:45:06.403734    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:06 crc kubenswrapper[4824]: I0122 09:45:06.403807    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:06 crc kubenswrapper[4824]: I0122 09:45:06.404506    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:06 crc kubenswrapper[4824]: E0122 09:45:06.404621    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447"
Jan 22 09:45:06 crc kubenswrapper[4824]: E0122 09:45:06.404838    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5"
Jan 22 09:45:06 crc kubenswrapper[4824]: E0122 09:45:06.404928    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xxtjz" podUID="f43b4618-291a-40d3-b1ca-32abfadaf376"
Jan 22 09:45:06 crc kubenswrapper[4824]: E0122 09:45:06.404982    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.404136    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.404175    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.404217    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.404376    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.408223    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.408570    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.413105    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.413279    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.413718    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c"
Jan 22 09:45:08 crc kubenswrapper[4824]: I0122 09:45:08.414877    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert"
Jan 22 09:45:13 crc kubenswrapper[4824]: I0122 09:45:13.759579    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.102249    4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.141409    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.142066    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.142340    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.143466    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.148604    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.148849    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.149044    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.149211    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.158082    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.158182    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.158199    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.158524    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.158615    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.158652    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.159230    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.159230    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.160028    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.161758    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.162330    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.163291    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.163962    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.165076    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.165420    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.168608    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xk8qh"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.169089    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.169168    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.169613    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.171496    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8878r"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.171898    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.176511    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.177225    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.177747    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.178102    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.181235    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.181598    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.181856    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.181983    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.182204    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.182700    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183956    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wrcr2"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.194812    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.182932    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.182989    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183051    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183109    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183164    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183217    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183271    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183324    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183407    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.183811    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.202925    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-822xl"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.205657    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.206001    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.206163    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.206384    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.206502    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.206687    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.206929    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.207055    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.207175    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.213618    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.213710    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.214022    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.214636    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.214973    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.215205    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.215646    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.215713    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.215862    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216062    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216097    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216240    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216388    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216447    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216570    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216623    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216572    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.216764    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.217002    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.217091    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.217747    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6m2np"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.218059    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.218155    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.218620    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.218072    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.219613    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221677    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221835    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-etcd-client\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221871    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-serving-cert\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221908    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-audit-policies\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221958    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221991    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb655775-5215-4e3c-856b-13a4c678582b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lw2st\" (UID: \"eb655775-5215-4e3c-856b-13a4c678582b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.222010    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.222030    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmjc5\" (UniqueName: \"kubernetes.io/projected/eb655775-5215-4e3c-856b-13a4c678582b-kube-api-access-gmjc5\") pod \"cluster-samples-operator-665b6dd947-lw2st\" (UID: \"eb655775-5215-4e3c-856b-13a4c678582b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.222046    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fba4ca2c-4f42-405d-8fc1-60a857973471-audit-dir\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.222067    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-encryption-config\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.222085    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ddls\" (UniqueName: \"kubernetes.io/projected/fba4ca2c-4f42-405d-8fc1-60a857973471-kube-api-access-9ddls\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221959    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.221698    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.223801    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.224219    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.224919    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.225294    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.225473    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.225624    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.225780    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.225960    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-b8244"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.226524    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.227332    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f79tb"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.228167    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mbh2m"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.228820    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.229626    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.249201    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.251407    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.252692    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbdgw"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.253258    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.253709    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.254261    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.254387    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.254267    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.254711    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255202    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255432    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255602    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255632    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255665    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255898    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.255941    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256041    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256126    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256192    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256222    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256673    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256780    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256833    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256918    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.256945    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257016    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257076    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257130    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257155    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257280    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257439    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257508    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257836    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.258144    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.258847    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257448    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.266701    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.267178    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.267425    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.275732    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.276547    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.279486    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.279623    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.257480    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.281047    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.282052    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.311877    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.312321    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.314952    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7qjxk"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.315561    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.316125    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rmsj9"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.316191    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.316644    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.316745    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.317236    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.317821    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.318180    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.318200    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.319145    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.319631    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x7psz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.320233    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.321516    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.322327    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323017    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-etcd-client\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323120    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-service-ca\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323201    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f5cc342-cace-4433-8fcb-25578dacdb68-serving-cert\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323282    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-encryption-config\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323396    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45dwl\" (UniqueName: \"kubernetes.io/projected/a2b6bacc-9256-4731-a61c-bbcb743aafe8-kube-api-access-45dwl\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323487    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/602843d5-8e62-4834-a7ae-af01d9f14644-serving-cert\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323580    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-serving-cert\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323651    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf09d978-94c1-460e-b1ee-29710101ac02-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323745    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-audit\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323837    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-serving-cert\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323933    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrv7\" (UniqueName: \"kubernetes.io/projected/bf09d978-94c1-460e-b1ee-29710101ac02-kube-api-access-lrrv7\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324038    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd56c\" (UniqueName: \"kubernetes.io/projected/602843d5-8e62-4834-a7ae-af01d9f14644-kube-api-access-cd56c\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324145    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.323953    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324332    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-config\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324462    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb655775-5215-4e3c-856b-13a4c678582b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lw2st\" (UID: \"eb655775-5215-4e3c-856b-13a4c678582b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324576    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9adf37cf-0b32-429f-ad06-819052944ee0-node-pullsecrets\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324679    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324778    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325242    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325481    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.324782    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/989e489d-6868-4601-82b0-71ee3474a4c1-serving-cert\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325646    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7tnh\" (UniqueName: \"kubernetes.io/projected/b5902322-e067-4ed7-967d-f71553df3532-kube-api-access-s7tnh\") pod \"control-plane-machine-set-operator-78cbb6b69f-zf5hj\" (UID: \"b5902322-e067-4ed7-967d-f71553df3532\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325669    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-oauth-serving-cert\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325704    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-encryption-config\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325731    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa-metrics-tls\") pod \"dns-operator-744455d44c-f79tb\" (UID: \"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa\") " pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325752    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tfb8\" (UniqueName: \"kubernetes.io/projected/cfe981a6-ed1d-411f-801d-b890b544ee5a-kube-api-access-5tfb8\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325789    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-ca\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325814    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5902322-e067-4ed7-967d-f71553df3532-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zf5hj\" (UID: \"b5902322-e067-4ed7-967d-f71553df3532\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325829    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325848    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-etcd-client\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325899    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-serving-cert\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325925    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9adf37cf-0b32-429f-ad06-819052944ee0-audit-dir\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325947    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fwzm\" (UniqueName: \"kubernetes.io/projected/989e489d-6868-4601-82b0-71ee3474a4c1-kube-api-access-8fwzm\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325981    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-audit-policies\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326002    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/989e489d-6868-4601-82b0-71ee3474a4c1-trusted-ca\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326029    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-client\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326053    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-service-ca\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326074    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/7f5cc342-cace-4433-8fcb-25578dacdb68-kube-api-access-vchhk\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326100    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-config\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326123    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-trusted-ca-bundle\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326167    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-927xj\" (UniqueName: \"kubernetes.io/projected/7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa-kube-api-access-927xj\") pod \"dns-operator-744455d44c-f79tb\" (UID: \"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa\") " pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326190    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-oauth-config\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326218    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/602843d5-8e62-4834-a7ae-af01d9f14644-available-featuregates\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326242    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-config\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326249    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326263    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-image-import-ca\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326283    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989e489d-6868-4601-82b0-71ee3474a4c1-config\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326312    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326335    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf09d978-94c1-460e-b1ee-29710101ac02-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326376    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-etcd-serving-ca\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326401    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2b6bacc-9256-4731-a61c-bbcb743aafe8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326427    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmjc5\" (UniqueName: \"kubernetes.io/projected/eb655775-5215-4e3c-856b-13a4c678582b-kube-api-access-gmjc5\") pod \"cluster-samples-operator-665b6dd947-lw2st\" (UID: \"eb655775-5215-4e3c-856b-13a4c678582b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326511    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fba4ca2c-4f42-405d-8fc1-60a857973471-audit-dir\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326539    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l2fj\" (UniqueName: \"kubernetes.io/projected/9adf37cf-0b32-429f-ad06-819052944ee0-kube-api-access-5l2fj\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326562    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b6bacc-9256-4731-a61c-bbcb743aafe8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326590    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ddls\" (UniqueName: \"kubernetes.io/projected/fba4ca2c-4f42-405d-8fc1-60a857973471-kube-api-access-9ddls\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.325664    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.327120    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.327295    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fba4ca2c-4f42-405d-8fc1-60a857973471-audit-dir\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.327699    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.329114    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.326074    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.329885    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fba4ca2c-4f42-405d-8fc1-60a857973471-audit-policies\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.332051    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-etcd-client\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.332383    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.332574    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-serving-cert\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.332654    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fba4ca2c-4f42-405d-8fc1-60a857973471-encryption-config\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.334142    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.334625    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.335114    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.335751    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.337340    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lprcb"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.337760    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.338736    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.339542    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4wqpq"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.340537    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.340544    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.341060    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb655775-5215-4e3c-856b-13a4c678582b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lw2st\" (UID: \"eb655775-5215-4e3c-856b-13a4c678582b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.341920    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xk8qh"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.343787    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.344144    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.345777    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.346978    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.349015    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.350975    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.352082    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f79tb"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.364725    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.364931    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.365477    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6m2np"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.372057    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.372131    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.372148    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rmsj9"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.375120    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.376377    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wrcr2"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.378428    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.380605    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.382028    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.383154    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.384182    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.384727    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8878r"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.385401    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.387600    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.388750    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mbh2m"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.389933    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.390971    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7qjxk"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.391973    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-822xl"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.392983    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.394011    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbdgw"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.396432    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.397592    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hmgmd"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.399025    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qdbht"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.399635    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.399857    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x7psz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.400216    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.401542    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.402943    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.404821    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410003    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410041    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lprcb"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410051    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qdbht"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410059    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hmgmd"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410068    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410077    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.410633    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zcblz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.411661    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.411719    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zcblz"]
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.424246    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427187    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9adf37cf-0b32-429f-ad06-819052944ee0-audit-dir\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427303    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9adf37cf-0b32-429f-ad06-819052944ee0-audit-dir\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427304    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fwzm\" (UniqueName: \"kubernetes.io/projected/989e489d-6868-4601-82b0-71ee3474a4c1-kube-api-access-8fwzm\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427389    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/989e489d-6868-4601-82b0-71ee3474a4c1-trusted-ca\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427415    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-client\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427440    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-service-ca\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427462    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-config\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427486    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-trusted-ca-bundle\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427506    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/7f5cc342-cace-4433-8fcb-25578dacdb68-kube-api-access-vchhk\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427522    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-oauth-config\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427552    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-927xj\" (UniqueName: \"kubernetes.io/projected/7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa-kube-api-access-927xj\") pod \"dns-operator-744455d44c-f79tb\" (UID: \"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa\") " pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427574    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/602843d5-8e62-4834-a7ae-af01d9f14644-available-featuregates\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427599    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-config\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427618    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989e489d-6868-4601-82b0-71ee3474a4c1-config\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427639    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-image-import-ca\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427662    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf09d978-94c1-460e-b1ee-29710101ac02-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427684    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-etcd-serving-ca\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427706    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2b6bacc-9256-4731-a61c-bbcb743aafe8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427725    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l2fj\" (UniqueName: \"kubernetes.io/projected/9adf37cf-0b32-429f-ad06-819052944ee0-kube-api-access-5l2fj\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427749    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b6bacc-9256-4731-a61c-bbcb743aafe8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427791    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-service-ca\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427812    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-etcd-client\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427835    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f5cc342-cace-4433-8fcb-25578dacdb68-serving-cert\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427856    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-encryption-config\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427878    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45dwl\" (UniqueName: \"kubernetes.io/projected/a2b6bacc-9256-4731-a61c-bbcb743aafe8-kube-api-access-45dwl\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427934    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/602843d5-8e62-4834-a7ae-af01d9f14644-serving-cert\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427956    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-audit\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427975    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-serving-cert\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.427995    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-serving-cert\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428017    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf09d978-94c1-460e-b1ee-29710101ac02-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428039    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrv7\" (UniqueName: \"kubernetes.io/projected/bf09d978-94c1-460e-b1ee-29710101ac02-kube-api-access-lrrv7\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428060    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd56c\" (UniqueName: \"kubernetes.io/projected/602843d5-8e62-4834-a7ae-af01d9f14644-kube-api-access-cd56c\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428089    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-config\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428112    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9adf37cf-0b32-429f-ad06-819052944ee0-node-pullsecrets\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428132    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428151    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/989e489d-6868-4601-82b0-71ee3474a4c1-serving-cert\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428175    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7tnh\" (UniqueName: \"kubernetes.io/projected/b5902322-e067-4ed7-967d-f71553df3532-kube-api-access-s7tnh\") pod \"control-plane-machine-set-operator-78cbb6b69f-zf5hj\" (UID: \"b5902322-e067-4ed7-967d-f71553df3532\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428198    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-oauth-serving-cert\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428223    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa-metrics-tls\") pod \"dns-operator-744455d44c-f79tb\" (UID: \"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa\") " pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428246    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tfb8\" (UniqueName: \"kubernetes.io/projected/cfe981a6-ed1d-411f-801d-b890b544ee5a-kube-api-access-5tfb8\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428269    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5902322-e067-4ed7-967d-f71553df3532-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zf5hj\" (UID: \"b5902322-e067-4ed7-967d-f71553df3532\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428291    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-ca\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.428916    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/989e489d-6868-4601-82b0-71ee3474a4c1-trusted-ca\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.429024    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-ca\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.429573    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-config\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.429636    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9adf37cf-0b32-429f-ad06-819052944ee0-node-pullsecrets\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.429746    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-config\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.430324    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-config\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.430464    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-trusted-ca-bundle\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.431638    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf09d978-94c1-460e-b1ee-29710101ac02-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.431832    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/602843d5-8e62-4834-a7ae-af01d9f14644-available-featuregates\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.432076    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-oauth-serving-cert\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.432097    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b6bacc-9256-4731-a61c-bbcb743aafe8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.432203    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-service-ca\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.432554    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/602843d5-8e62-4834-a7ae-af01d9f14644-serving-cert\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.432810    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/989e489d-6868-4601-82b0-71ee3474a4c1-config\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.433055    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-oauth-config\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.433247    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-service-ca\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.433697    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2b6bacc-9256-4731-a61c-bbcb743aafe8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.434292    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-serving-cert\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.434456    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f5cc342-cace-4433-8fcb-25578dacdb68-serving-cert\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.435489    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-serving-cert\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.436864    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7f5cc342-cace-4433-8fcb-25578dacdb68-etcd-client\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.437393    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf09d978-94c1-460e-b1ee-29710101ac02-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.438207    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/989e489d-6868-4601-82b0-71ee3474a4c1-serving-cert\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.444005    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.452541    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-encryption-config\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.463805    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.475120    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9adf37cf-0b32-429f-ad06-819052944ee0-etcd-client\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.488183    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.504101    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.523839    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.531174    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-audit\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.544683    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.553060    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-etcd-serving-ca\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.564008    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.573690    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-image-import-ca\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.590173    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.600712    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9adf37cf-0b32-429f-ad06-819052944ee0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.603413    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.623829    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.635143    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa-metrics-tls\") pod \"dns-operator-744455d44c-f79tb\" (UID: \"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa\") " pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.644758    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.664486    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.684224    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.704269    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.724167    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.744631    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.763480    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.783443    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.804963    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.824920    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.844564    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.864581    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.884317    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.905131    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.925329    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.936011    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b5902322-e067-4ed7-967d-f71553df3532-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zf5hj\" (UID: \"b5902322-e067-4ed7-967d-f71553df3532\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.944301    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt"
Jan 22 09:45:15 crc kubenswrapper[4824]: I0122 09:45:15.984795    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.004581    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.024720    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.043880    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.065230    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.084439    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.104244    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.125026    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.144907    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.165666    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.204384    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.224602    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.244773    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.275153    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.283926    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.305619    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.322821    4824 request.go:700] Waited for 1.004776886s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.324215    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.346881    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.364791    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.384301    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.404784    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.424470    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.445098    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.465091    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.484573    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.505097    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.524405    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.544687    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.565286    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.584099    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.604788    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.652875    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmjc5\" (UniqueName: \"kubernetes.io/projected/eb655775-5215-4e3c-856b-13a4c678582b-kube-api-access-gmjc5\") pod \"cluster-samples-operator-665b6dd947-lw2st\" (UID: \"eb655775-5215-4e3c-856b-13a4c678582b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.664434    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.674439    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ddls\" (UniqueName: \"kubernetes.io/projected/fba4ca2c-4f42-405d-8fc1-60a857973471-kube-api-access-9ddls\") pod \"apiserver-7bbb656c7d-qlzt7\" (UID: \"fba4ca2c-4f42-405d-8fc1-60a857973471\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.684944    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.693647    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.703396    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.724757    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.744530    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.764783    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.785506    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.805328    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.824499    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.844476    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.863983    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.868472    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st"]
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.884899    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.904179    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.924517    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.944437    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.964077    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.967646    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:16 crc kubenswrapper[4824]: I0122 09:45:16.984103    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.004384    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.024107    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.044798    4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.063966    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.084252    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.104691    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.124379    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.133626    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"]
Jan 22 09:45:17 crc kubenswrapper[4824]: W0122 09:45:17.142017    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfba4ca2c_4f42_405d_8fc1_60a857973471.slice/crio-13e6ca8e26a07b3f940ba577797c00f706c47b06e8fc127efaf919c7388a4e9f WatchSource:0}: Error finding container 13e6ca8e26a07b3f940ba577797c00f706c47b06e8fc127efaf919c7388a4e9f: Status 404 returned error can't find the container with id 13e6ca8e26a07b3f940ba577797c00f706c47b06e8fc127efaf919c7388a4e9f
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.144334    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.165096    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.184725    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.204872    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.225436    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.261974    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fwzm\" (UniqueName: \"kubernetes.io/projected/989e489d-6868-4601-82b0-71ee3474a4c1-kube-api-access-8fwzm\") pod \"console-operator-58897d9998-wrcr2\" (UID: \"989e489d-6868-4601-82b0-71ee3474a4c1\") " pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.282969    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd56c\" (UniqueName: \"kubernetes.io/projected/602843d5-8e62-4834-a7ae-af01d9f14644-kube-api-access-cd56c\") pod \"openshift-config-operator-7777fb866f-b5fvk\" (UID: \"602843d5-8e62-4834-a7ae-af01d9f14644\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.303252    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45dwl\" (UniqueName: \"kubernetes.io/projected/a2b6bacc-9256-4731-a61c-bbcb743aafe8-kube-api-access-45dwl\") pod \"openshift-apiserver-operator-796bbdcf4f-6vpgm\" (UID: \"a2b6bacc-9256-4731-a61c-bbcb743aafe8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.321328    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vchhk\" (UniqueName: \"kubernetes.io/projected/7f5cc342-cace-4433-8fcb-25578dacdb68-kube-api-access-vchhk\") pod \"etcd-operator-b45778765-6m2np\" (UID: \"7f5cc342-cace-4433-8fcb-25578dacdb68\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.323005    4824 request.go:700] Waited for 1.89251787s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/serviceaccounts/dns-operator/token
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.339235    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-927xj\" (UniqueName: \"kubernetes.io/projected/7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa-kube-api-access-927xj\") pod \"dns-operator-744455d44c-f79tb\" (UID: \"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa\") " pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.363845    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrv7\" (UniqueName: \"kubernetes.io/projected/bf09d978-94c1-460e-b1ee-29710101ac02-kube-api-access-lrrv7\") pod \"openshift-controller-manager-operator-756b6f6bc6-zkj26\" (UID: \"bf09d978-94c1-460e-b1ee-29710101ac02\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.377183    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l2fj\" (UniqueName: \"kubernetes.io/projected/9adf37cf-0b32-429f-ad06-819052944ee0-kube-api-access-5l2fj\") pod \"apiserver-76f77b778f-mbh2m\" (UID: \"9adf37cf-0b32-429f-ad06-819052944ee0\") " pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.397571    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7tnh\" (UniqueName: \"kubernetes.io/projected/b5902322-e067-4ed7-967d-f71553df3532-kube-api-access-s7tnh\") pod \"control-plane-machine-set-operator-78cbb6b69f-zf5hj\" (UID: \"b5902322-e067-4ed7-967d-f71553df3532\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.417634    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tfb8\" (UniqueName: \"kubernetes.io/projected/cfe981a6-ed1d-411f-801d-b890b544ee5a-kube-api-access-5tfb8\") pod \"console-f9d7485db-8878r\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") " pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.420415    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.431956    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455530    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455582    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt9x8\" (UniqueName: \"kubernetes.io/projected/fc660b59-eb41-4c18-84a4-ebd782752570-kube-api-access-zt9x8\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455605    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9n59\" (UniqueName: \"kubernetes.io/projected/78d695a9-430a-4540-8a70-5c2bcfc40b34-kube-api-access-p9n59\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455633    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9633603d-c8b9-4f94-a3c5-3d882308a586-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455657    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxkl2\" (UniqueName: \"kubernetes.io/projected/28370e19-78ca-49e9-916f-aaa80793d644-kube-api-access-cxkl2\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455694    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-metrics-certs\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455736    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-stats-auth\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455765    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78d695a9-430a-4540-8a70-5c2bcfc40b34-serving-cert\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455792    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-default-certificate\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455816    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455838    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-auth-proxy-config\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455861    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-service-ca-bundle\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455899    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-client-ca\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455921    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455944    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526c6812-a825-4d53-90fb-fa01195fa33a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455968    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxjwz\" (UniqueName: \"kubernetes.io/projected/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-kube-api-access-rxjwz\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.455990    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456014    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95a827d-6910-45c6-8dbf-a45a4cc6578f-config\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456044    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456065    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-serving-cert\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456086    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456145    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28370e19-78ca-49e9-916f-aaa80793d644-auth-proxy-config\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456166    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-client-ca\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456189    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456211    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbj7x\" (UniqueName: \"kubernetes.io/projected/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-kube-api-access-dbj7x\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456232    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1215925-7f63-449c-8aa1-879ac01ed726-bound-sa-token\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456274    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456298    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456322    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp4bn\" (UniqueName: \"kubernetes.io/projected/a1215925-7f63-449c-8aa1-879ac01ed726-kube-api-access-jp4bn\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456393    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-audit-policies\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456421    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/28370e19-78ca-49e9-916f-aaa80793d644-images\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456447    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plfc8\" (UniqueName: \"kubernetes.io/projected/697490dd-5ba8-4719-b8a3-87a9b9ed08b3-kube-api-access-plfc8\") pod \"migrator-59844c95c7-q4jqv\" (UID: \"697490dd-5ba8-4719-b8a3-87a9b9ed08b3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456500    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456862    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-config\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456928    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-config\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456964    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-machine-approver-tls\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.456989    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.457010    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dcf185-540b-4a55-89e6-cbc3b904a15c-service-ca-bundle\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.457059    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b95a827d-6910-45c6-8dbf-a45a4cc6578f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.457088    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28370e19-78ca-49e9-916f-aaa80793d644-proxy-tls\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.457136    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-config\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.457220    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37205799-5637-4c12-8215-16f21fe22c72-serving-cert\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.457250    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7hcs\" (UniqueName: \"kubernetes.io/projected/12dcf185-540b-4a55-89e6-cbc3b904a15c-kube-api-access-b7hcs\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459594    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dns\" (UniqueName: \"kubernetes.io/projected/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-kube-api-access-d4dns\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459665    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fc660b59-eb41-4c18-84a4-ebd782752570-audit-dir\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459724    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459758    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/526c6812-a825-4d53-90fb-fa01195fa33a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459837    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9633603d-c8b9-4f94-a3c5-3d882308a586-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459870    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-bound-sa-token\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459908    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.459931    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2zc4\" (UniqueName: \"kubernetes.io/projected/37205799-5637-4c12-8215-16f21fe22c72-kube-api-access-x2zc4\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460084    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-tls\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460131    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526c6812-a825-4d53-90fb-fa01195fa33a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460151    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460199    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460256    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460282    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-trusted-ca\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460319    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-config\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460344    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.460702    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:17.960687301 +0000 UTC m=+150.864164354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.460383    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1215925-7f63-449c-8aa1-879ac01ed726-trusted-ca\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.461029    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tv7t\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-kube-api-access-7tv7t\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.461054    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b95a827d-6910-45c6-8dbf-a45a4cc6578f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.461088    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1215925-7f63-449c-8aa1-879ac01ed726-metrics-tls\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.461158    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-certificates\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.469037    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.481196    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.492609    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.499506    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.562930    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563122    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526c6812-a825-4d53-90fb-fa01195fa33a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563195    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563225    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28ae5f2e-c8dd-4d2c-b318-9677541a6656-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563268    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ssf7\" (UniqueName: \"kubernetes.io/projected/eae9e020-d1a0-467f-b10b-9e022f357fc3-kube-api-access-4ssf7\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563346    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-mountpoint-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563389    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43bcec80-1438-4a3f-ba1c-48b2055751ac-config\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563426    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1215925-7f63-449c-8aa1-879ac01ed726-trusted-ca\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563454    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tv7t\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-kube-api-access-7tv7t\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563487    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-certificates\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563509    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9n59\" (UniqueName: \"kubernetes.io/projected/78d695a9-430a-4540-8a70-5c2bcfc40b34-kube-api-access-p9n59\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563535    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skjdt\" (UniqueName: \"kubernetes.io/projected/43bcec80-1438-4a3f-ba1c-48b2055751ac-kube-api-access-skjdt\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563567    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563603    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxkl2\" (UniqueName: \"kubernetes.io/projected/28370e19-78ca-49e9-916f-aaa80793d644-kube-api-access-cxkl2\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563654    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-stats-auth\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.563697    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.063664208 +0000 UTC m=+150.967141231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563742    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78d695a9-430a-4540-8a70-5c2bcfc40b34-serving-cert\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563799    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563834    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lfjz\" (UniqueName: \"kubernetes.io/projected/cb724a14-a04d-458a-8350-65e8f9d5a69e-kube-api-access-5lfjz\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563860    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-default-certificate\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563886    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-auth-proxy-config\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563912    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-service-ca-bundle\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563936    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-socket-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.563959    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564000    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-client-ca\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564025    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526c6812-a825-4d53-90fb-fa01195fa33a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564049    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxjwz\" (UniqueName: \"kubernetes.io/projected/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-kube-api-access-rxjwz\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564073    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564116    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564157    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564181    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564204    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbj7x\" (UniqueName: \"kubernetes.io/projected/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-kube-api-access-dbj7x\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564230    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h62q\" (UniqueName: \"kubernetes.io/projected/7747087d-a669-4f55-978c-5bce681a7e2e-kube-api-access-9h62q\") pod \"ingress-canary-qdbht\" (UID: \"7747087d-a669-4f55-978c-5bce681a7e2e\") " pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564257    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b48m\" (UniqueName: \"kubernetes.io/projected/c8b117b7-e048-4a4a-9592-1d3eb821d309-kube-api-access-5b48m\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564280    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/faa5f086-812d-4c56-9491-fd7147a6706a-profile-collector-cert\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564314    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlc4\" (UniqueName: \"kubernetes.io/projected/28ae5f2e-c8dd-4d2c-b318-9677541a6656-kube-api-access-ltlc4\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564383    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-audit-policies\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564409    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp4bn\" (UniqueName: \"kubernetes.io/projected/a1215925-7f63-449c-8aa1-879ac01ed726-kube-api-access-jp4bn\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564429    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plfc8\" (UniqueName: \"kubernetes.io/projected/697490dd-5ba8-4719-b8a3-87a9b9ed08b3-kube-api-access-plfc8\") pod \"migrator-59844c95c7-q4jqv\" (UID: \"697490dd-5ba8-4719-b8a3-87a9b9ed08b3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564454    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564483    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0e085a8e-5760-4859-bf6b-ac4fac6aa9cf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hpp2x\" (UID: \"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564502    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5ad14705-3767-49da-86e9-544b5259b066-srv-cert\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564537    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-config\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564559    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43bcec80-1438-4a3f-ba1c-48b2055751ac-serving-cert\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564580    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-config\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564604    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-config\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564628    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46zr8\" (UniqueName: \"kubernetes.io/projected/5ad14705-3767-49da-86e9-544b5259b066-kube-api-access-46zr8\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564648    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/faa5f086-812d-4c56-9491-fd7147a6706a-srv-cert\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564689    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-images\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564719    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-machine-approver-tls\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564743    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564773    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ae5f2e-c8dd-4d2c-b318-9677541a6656-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564811    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28370e19-78ca-49e9-916f-aaa80793d644-proxy-tls\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564833    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dcf185-540b-4a55-89e6-cbc3b904a15c-service-ca-bundle\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564859    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97038c7c-c8e0-451d-86a5-93e5f36d5e33-webhook-cert\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564883    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-config\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564921    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29832635-7ac3-4472-afcd-3f06b7a2fe91-config-volume\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564945    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb724a14-a04d-458a-8350-65e8f9d5a69e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564969    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwkrp\" (UniqueName: \"kubernetes.io/projected/2463610b-364e-4154-83ee-b6c00a6c6da5-kube-api-access-zwkrp\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.564998    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fc660b59-eb41-4c18-84a4-ebd782752570-audit-dir\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565020    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6803cfcb-a020-4f9c-a668-a2df6abbb407-signing-cabundle\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565043    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x4jm\" (UniqueName: \"kubernetes.io/projected/3a616855-646b-46d9-a8b4-76130bd53772-kube-api-access-7x4jm\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565068    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/526c6812-a825-4d53-90fb-fa01195fa33a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565095    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9633603d-c8b9-4f94-a3c5-3d882308a586-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565115    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565138    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2zc4\" (UniqueName: \"kubernetes.io/projected/37205799-5637-4c12-8215-16f21fe22c72-kube-api-access-x2zc4\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565158    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-tls\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565181    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-plugins-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565205    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws546\" (UniqueName: \"kubernetes.io/projected/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-kube-api-access-ws546\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565232    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7596399-0f46-4c4a-97ae-379efd65e7af-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565277    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565300    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-trusted-ca\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565326    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565708    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-config\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565759    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565785    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz98s\" (UniqueName: \"kubernetes.io/projected/29832635-7ac3-4472-afcd-3f06b7a2fe91-kube-api-access-qz98s\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565807    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a616855-646b-46d9-a8b4-76130bd53772-node-bootstrap-token\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565833    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b95a827d-6910-45c6-8dbf-a45a4cc6578f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565855    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1215925-7f63-449c-8aa1-879ac01ed726-metrics-tls\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565879    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eae9e020-d1a0-467f-b10b-9e022f357fc3-secret-volume\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565900    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-csi-data-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565939    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt9x8\" (UniqueName: \"kubernetes.io/projected/fc660b59-eb41-4c18-84a4-ebd782752570-kube-api-access-zt9x8\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565970    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9633603d-c8b9-4f94-a3c5-3d882308a586-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.565993    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-metrics-certs\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566027    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2699m\" (UniqueName: \"kubernetes.io/projected/0e085a8e-5760-4859-bf6b-ac4fac6aa9cf-kube-api-access-2699m\") pod \"multus-admission-controller-857f4d67dd-hpp2x\" (UID: \"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566052    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7596399-0f46-4c4a-97ae-379efd65e7af-config\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566066    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566083    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566125    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6803cfcb-a020-4f9c-a668-a2df6abbb407-signing-key\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566152    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566183    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvtfl\" (UniqueName: \"kubernetes.io/projected/97038c7c-c8e0-451d-86a5-93e5f36d5e33-kube-api-access-tvtfl\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566209    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52kr2\" (UniqueName: \"kubernetes.io/projected/faa5f086-812d-4c56-9491-fd7147a6706a-kube-api-access-52kr2\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566239    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95a827d-6910-45c6-8dbf-a45a4cc6578f-config\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566265    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c2ebdfa-baca-4839-8524-e4161b7cbe89-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-877dz\" (UID: \"9c2ebdfa-baca-4839-8524-e4161b7cbe89\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566303    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7747087d-a669-4f55-978c-5bce681a7e2e-cert\") pod \"ingress-canary-qdbht\" (UID: \"7747087d-a669-4f55-978c-5bce681a7e2e\") " pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566329    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-registration-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.566989    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-stats-auth\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.567999    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-audit-policies\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.569388    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-config\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.571393    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.571425    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-config\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.572494    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/526c6812-a825-4d53-90fb-fa01195fa33a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.572661    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.573002    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fc660b59-eb41-4c18-84a4-ebd782752570-audit-dir\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.573035    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12dcf185-540b-4a55-89e6-cbc3b904a15c-service-ca-bundle\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.573057    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-client-ca\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575442    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-serving-cert\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575489    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28370e19-78ca-49e9-916f-aaa80793d644-auth-proxy-config\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575512    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-client-ca\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575553    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575575    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1215925-7f63-449c-8aa1-879ac01ed726-bound-sa-token\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575603    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7596399-0f46-4c4a-97ae-379efd65e7af-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575643    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/28370e19-78ca-49e9-916f-aaa80793d644-images\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575665    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29832635-7ac3-4472-afcd-3f06b7a2fe91-metrics-tls\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575688    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5ad14705-3767-49da-86e9-544b5259b066-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575715    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575765    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb724a14-a04d-458a-8350-65e8f9d5a69e-proxy-tls\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575792    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575843    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l49bn\" (UniqueName: \"kubernetes.io/projected/4809673b-3d53-4bd5-8739-9be6a2443e60-kube-api-access-l49bn\") pod \"downloads-7954f5f757-rmsj9\" (UID: \"4809673b-3d53-4bd5-8739-9be6a2443e60\") " pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575869    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b95a827d-6910-45c6-8dbf-a45a4cc6578f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575878    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575894    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62kg\" (UniqueName: \"kubernetes.io/projected/6803cfcb-a020-4f9c-a668-a2df6abbb407-kube-api-access-m62kg\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575918    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2b2t\" (UniqueName: \"kubernetes.io/projected/9c2ebdfa-baca-4839-8524-e4161b7cbe89-kube-api-access-b2b2t\") pod \"package-server-manager-789f6589d5-877dz\" (UID: \"9c2ebdfa-baca-4839-8524-e4161b7cbe89\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.575971    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a616855-646b-46d9-a8b4-76130bd53772-certs\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576039    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37205799-5637-4c12-8215-16f21fe22c72-serving-cert\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576064    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7hcs\" (UniqueName: \"kubernetes.io/projected/12dcf185-540b-4a55-89e6-cbc3b904a15c-kube-api-access-b7hcs\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576087    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eae9e020-d1a0-467f-b10b-9e022f357fc3-config-volume\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576139    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dns\" (UniqueName: \"kubernetes.io/projected/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-kube-api-access-d4dns\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576163    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576172    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-config\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576184    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/97038c7c-c8e0-451d-86a5-93e5f36d5e33-tmpfs\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576207    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97038c7c-c8e0-451d-86a5-93e5f36d5e33-apiservice-cert\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576215    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-certificates\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576234    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-bound-sa-token\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576396    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-default-certificate\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.576641    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1215925-7f63-449c-8aa1-879ac01ed726-trusted-ca\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.577089    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-auth-proxy-config\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.577329    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b95a827d-6910-45c6-8dbf-a45a4cc6578f-config\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.577565    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.577566    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/28370e19-78ca-49e9-916f-aaa80793d644-images\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.577869    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9633603d-c8b9-4f94-a3c5-3d882308a586-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.578881    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.07870593 +0000 UTC m=+150.982182923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.579861    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-trusted-ca\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.580509    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-machine-approver-tls\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.581230    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.581262    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.590435    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-client-ca\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.590904    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28370e19-78ca-49e9-916f-aaa80793d644-proxy-tls\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.591401    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.591751    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28370e19-78ca-49e9-916f-aaa80793d644-auth-proxy-config\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.592105    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-config\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.592909    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-tls\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.593163    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526c6812-a825-4d53-90fb-fa01195fa33a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.593334    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.593466    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-serving-cert\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.594430    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37205799-5637-4c12-8215-16f21fe22c72-service-ca-bundle\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.594819    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f79tb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.595021    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.602640    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b95a827d-6910-45c6-8dbf-a45a4cc6578f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.603327    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.604774    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.606932    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a1215925-7f63-449c-8aa1-879ac01ed726-metrics-tls\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.607701    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9633603d-c8b9-4f94-a3c5-3d882308a586-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.607765    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.607866    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.607895    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37205799-5637-4c12-8215-16f21fe22c72-serving-cert\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.608153    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78d695a9-430a-4540-8a70-5c2bcfc40b34-serving-cert\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.608263    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.609776    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12dcf185-540b-4a55-89e6-cbc3b904a15c-metrics-certs\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.611856    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.612285    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp4bn\" (UniqueName: \"kubernetes.io/projected/a1215925-7f63-449c-8aa1-879ac01ed726-kube-api-access-jp4bn\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.620347    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.621392    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.636215    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plfc8\" (UniqueName: \"kubernetes.io/projected/697490dd-5ba8-4719-b8a3-87a9b9ed08b3-kube-api-access-plfc8\") pod \"migrator-59844c95c7-q4jqv\" (UID: \"697490dd-5ba8-4719-b8a3-87a9b9ed08b3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.643264    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.673155    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tv7t\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-kube-api-access-7tv7t\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.677919    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.678011    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.177985998 +0000 UTC m=+151.081462991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678178    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28ae5f2e-c8dd-4d2c-b318-9677541a6656-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678219    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ssf7\" (UniqueName: \"kubernetes.io/projected/eae9e020-d1a0-467f-b10b-9e022f357fc3-kube-api-access-4ssf7\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678246    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-mountpoint-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678267    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43bcec80-1438-4a3f-ba1c-48b2055751ac-config\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678300    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skjdt\" (UniqueName: \"kubernetes.io/projected/43bcec80-1438-4a3f-ba1c-48b2055751ac-kube-api-access-skjdt\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678349    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lfjz\" (UniqueName: \"kubernetes.io/projected/cb724a14-a04d-458a-8350-65e8f9d5a69e-kube-api-access-5lfjz\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678390    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-socket-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678417    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678457    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h62q\" (UniqueName: \"kubernetes.io/projected/7747087d-a669-4f55-978c-5bce681a7e2e-kube-api-access-9h62q\") pod \"ingress-canary-qdbht\" (UID: \"7747087d-a669-4f55-978c-5bce681a7e2e\") " pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678477    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b48m\" (UniqueName: \"kubernetes.io/projected/c8b117b7-e048-4a4a-9592-1d3eb821d309-kube-api-access-5b48m\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678497    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/faa5f086-812d-4c56-9491-fd7147a6706a-profile-collector-cert\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678529    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlc4\" (UniqueName: \"kubernetes.io/projected/28ae5f2e-c8dd-4d2c-b318-9677541a6656-kube-api-access-ltlc4\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678555    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0e085a8e-5760-4859-bf6b-ac4fac6aa9cf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hpp2x\" (UID: \"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678576    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5ad14705-3767-49da-86e9-544b5259b066-srv-cert\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678582    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-mountpoint-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678944    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-socket-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.678597    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43bcec80-1438-4a3f-ba1c-48b2055751ac-serving-cert\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679058    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-config\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679110    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46zr8\" (UniqueName: \"kubernetes.io/projected/5ad14705-3767-49da-86e9-544b5259b066-kube-api-access-46zr8\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679155    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43bcec80-1438-4a3f-ba1c-48b2055751ac-config\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679301    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/faa5f086-812d-4c56-9491-fd7147a6706a-srv-cert\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679337    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-images\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679376    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ae5f2e-c8dd-4d2c-b318-9677541a6656-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679407    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97038c7c-c8e0-451d-86a5-93e5f36d5e33-webhook-cert\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679431    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29832635-7ac3-4472-afcd-3f06b7a2fe91-config-volume\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679451    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb724a14-a04d-458a-8350-65e8f9d5a69e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679474    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwkrp\" (UniqueName: \"kubernetes.io/projected/2463610b-364e-4154-83ee-b6c00a6c6da5-kube-api-access-zwkrp\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679497    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6803cfcb-a020-4f9c-a668-a2df6abbb407-signing-cabundle\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679520    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x4jm\" (UniqueName: \"kubernetes.io/projected/3a616855-646b-46d9-a8b4-76130bd53772-kube-api-access-7x4jm\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679563    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7596399-0f46-4c4a-97ae-379efd65e7af-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679585    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-plugins-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679613    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws546\" (UniqueName: \"kubernetes.io/projected/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-kube-api-access-ws546\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679644    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679673    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz98s\" (UniqueName: \"kubernetes.io/projected/29832635-7ac3-4472-afcd-3f06b7a2fe91-kube-api-access-qz98s\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679694    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a616855-646b-46d9-a8b4-76130bd53772-node-bootstrap-token\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679720    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eae9e020-d1a0-467f-b10b-9e022f357fc3-secret-volume\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679750    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-csi-data-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679788    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2699m\" (UniqueName: \"kubernetes.io/projected/0e085a8e-5760-4859-bf6b-ac4fac6aa9cf-kube-api-access-2699m\") pod \"multus-admission-controller-857f4d67dd-hpp2x\" (UID: \"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679811    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7596399-0f46-4c4a-97ae-379efd65e7af-config\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679840    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6803cfcb-a020-4f9c-a668-a2df6abbb407-signing-key\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679860    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679881    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvtfl\" (UniqueName: \"kubernetes.io/projected/97038c7c-c8e0-451d-86a5-93e5f36d5e33-kube-api-access-tvtfl\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679904    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c2ebdfa-baca-4839-8524-e4161b7cbe89-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-877dz\" (UID: \"9c2ebdfa-baca-4839-8524-e4161b7cbe89\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679928    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52kr2\" (UniqueName: \"kubernetes.io/projected/faa5f086-812d-4c56-9491-fd7147a6706a-kube-api-access-52kr2\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679951    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7747087d-a669-4f55-978c-5bce681a7e2e-cert\") pod \"ingress-canary-qdbht\" (UID: \"7747087d-a669-4f55-978c-5bce681a7e2e\") " pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.679972    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-registration-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680018    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7596399-0f46-4c4a-97ae-379efd65e7af-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680040    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29832635-7ac3-4472-afcd-3f06b7a2fe91-metrics-tls\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680064    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5ad14705-3767-49da-86e9-544b5259b066-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680089    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb724a14-a04d-458a-8350-65e8f9d5a69e-proxy-tls\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680112    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680135    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l49bn\" (UniqueName: \"kubernetes.io/projected/4809673b-3d53-4bd5-8739-9be6a2443e60-kube-api-access-l49bn\") pod \"downloads-7954f5f757-rmsj9\" (UID: \"4809673b-3d53-4bd5-8739-9be6a2443e60\") " pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680164    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62kg\" (UniqueName: \"kubernetes.io/projected/6803cfcb-a020-4f9c-a668-a2df6abbb407-kube-api-access-m62kg\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680185    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2b2t\" (UniqueName: \"kubernetes.io/projected/9c2ebdfa-baca-4839-8524-e4161b7cbe89-kube-api-access-b2b2t\") pod \"package-server-manager-789f6589d5-877dz\" (UID: \"9c2ebdfa-baca-4839-8524-e4161b7cbe89\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680206    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a616855-646b-46d9-a8b4-76130bd53772-certs\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680243    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eae9e020-d1a0-467f-b10b-9e022f357fc3-config-volume\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680274    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/97038c7c-c8e0-451d-86a5-93e5f36d5e33-tmpfs\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.680312    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97038c7c-c8e0-451d-86a5-93e5f36d5e33-apiservice-cert\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.681314    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-config\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.682872    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.18285032 +0000 UTC m=+151.086327313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.683169    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-images\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.683243    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97038c7c-c8e0-451d-86a5-93e5f36d5e33-apiservice-cert\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.683426    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-plugins-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.683931    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/526c6812-a825-4d53-90fb-fa01195fa33a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tkrmz\" (UID: \"526c6812-a825-4d53-90fb-fa01195fa33a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.684151    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/faa5f086-812d-4c56-9491-fd7147a6706a-srv-cert\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.684248    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/29832635-7ac3-4472-afcd-3f06b7a2fe91-metrics-tls\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.684447    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0e085a8e-5760-4859-bf6b-ac4fac6aa9cf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-hpp2x\" (UID: \"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.684730    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/43bcec80-1438-4a3f-ba1c-48b2055751ac-serving-cert\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685149    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb724a14-a04d-458a-8350-65e8f9d5a69e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685212    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ae5f2e-c8dd-4d2c-b318-9677541a6656-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685310    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eae9e020-d1a0-467f-b10b-9e022f357fc3-config-volume\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685348    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/97038c7c-c8e0-451d-86a5-93e5f36d5e33-tmpfs\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685505    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-csi-data-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685533    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c8b117b7-e048-4a4a-9592-1d3eb821d309-registration-dir\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.685825    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5ad14705-3767-49da-86e9-544b5259b066-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.686346    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6803cfcb-a020-4f9c-a668-a2df6abbb407-signing-cabundle\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.686824    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.687034    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/29832635-7ac3-4472-afcd-3f06b7a2fe91-config-volume\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.688509    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7596399-0f46-4c4a-97ae-379efd65e7af-config\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.696326    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/faa5f086-812d-4c56-9491-fd7147a6706a-profile-collector-cert\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.696412    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.698417    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3a616855-646b-46d9-a8b4-76130bd53772-certs\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.698662    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8878r"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.698850    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3a616855-646b-46d9-a8b4-76130bd53772-node-bootstrap-token\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.700242    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eae9e020-d1a0-467f-b10b-9e022f357fc3-secret-volume\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.700281    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5ad14705-3767-49da-86e9-544b5259b066-srv-cert\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.700704    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7596399-0f46-4c4a-97ae-379efd65e7af-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.700766    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97038c7c-c8e0-451d-86a5-93e5f36d5e33-webhook-cert\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.700905    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/9c2ebdfa-baca-4839-8524-e4161b7cbe89-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-877dz\" (UID: \"9c2ebdfa-baca-4839-8524-e4161b7cbe89\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.701562    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb724a14-a04d-458a-8350-65e8f9d5a69e-proxy-tls\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.701957    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28ae5f2e-c8dd-4d2c-b318-9677541a6656-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.701959    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6803cfcb-a020-4f9c-a668-a2df6abbb407-signing-key\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.703586    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7747087d-a669-4f55-978c-5bce681a7e2e-cert\") pod \"ingress-canary-qdbht\" (UID: \"7747087d-a669-4f55-978c-5bce681a7e2e\") " pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.705259    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:17 crc kubenswrapper[4824]: W0122 09:45:17.709438    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfe981a6_ed1d_411f_801d_b890b544ee5a.slice/crio-d3d480caa0c4757f8d4123d3c8684149691d7c5ccdbe6205c6655e4b7402478a WatchSource:0}: Error finding container d3d480caa0c4757f8d4123d3c8684149691d7c5ccdbe6205c6655e4b7402478a: Status 404 returned error can't find the container with id d3d480caa0c4757f8d4123d3c8684149691d7c5ccdbe6205c6655e4b7402478a
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.721334    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxjwz\" (UniqueName: \"kubernetes.io/projected/4533ce52-cfcf-4df1-9f6f-e8d6febd7719-kube-api-access-rxjwz\") pod \"machine-approver-56656f9798-zg6fj\" (UID: \"4533ce52-cfcf-4df1-9f6f-e8d6febd7719\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.747045    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxkl2\" (UniqueName: \"kubernetes.io/projected/28370e19-78ca-49e9-916f-aaa80793d644-kube-api-access-cxkl2\") pod \"machine-config-operator-74547568cd-97mmf\" (UID: \"28370e19-78ca-49e9-916f-aaa80793d644\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.782879    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.782911    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9n59\" (UniqueName: \"kubernetes.io/projected/78d695a9-430a-4540-8a70-5c2bcfc40b34-kube-api-access-p9n59\") pod \"route-controller-manager-6576b87f9c-5n9hc\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.783239    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.783440    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.283412187 +0000 UTC m=+151.186889180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.783863    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.784285    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.284261088 +0000 UTC m=+151.187738111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.788122    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1215925-7f63-449c-8aa1-879ac01ed726-bound-sa-token\") pod \"ingress-operator-5b745b69d9-29d5x\" (UID: \"a1215925-7f63-449c-8aa1-879ac01ed726\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.799512    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dns\" (UniqueName: \"kubernetes.io/projected/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-kube-api-access-d4dns\") pod \"controller-manager-879f6c89f-xk8qh\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.825111    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wrcr2"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.830036    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6m2np"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.832257    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2zc4\" (UniqueName: \"kubernetes.io/projected/37205799-5637-4c12-8215-16f21fe22c72-kube-api-access-x2zc4\") pod \"authentication-operator-69f744f599-l9tjj\" (UID: \"37205799-5637-4c12-8215-16f21fe22c72\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.832577    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"
Jan 22 09:45:17 crc kubenswrapper[4824]: W0122 09:45:17.836722    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod989e489d_6868_4601_82b0_71ee3474a4c1.slice/crio-49c6343218dd1ccac2218b74d9eebc52bc1b0da36db71b780a6480d8215a3352 WatchSource:0}: Error finding container 49c6343218dd1ccac2218b74d9eebc52bc1b0da36db71b780a6480d8215a3352: Status 404 returned error can't find the container with id 49c6343218dd1ccac2218b74d9eebc52bc1b0da36db71b780a6480d8215a3352
Jan 22 09:45:17 crc kubenswrapper[4824]: W0122 09:45:17.840231    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f5cc342_cace_4433_8fcb_25578dacdb68.slice/crio-bd90fda2cc9359c86d25d176dff8722f7a8980e0fb51e32aadf9c4349cf78c81 WatchSource:0}: Error finding container bd90fda2cc9359c86d25d176dff8722f7a8980e0fb51e32aadf9c4349cf78c81: Status 404 returned error can't find the container with id bd90fda2cc9359c86d25d176dff8722f7a8980e0fb51e32aadf9c4349cf78c81
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.842825    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt9x8\" (UniqueName: \"kubernetes.io/projected/fc660b59-eb41-4c18-84a4-ebd782752570-kube-api-access-zt9x8\") pod \"oauth-openshift-558db77b4-822xl\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") " pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.859539    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7hcs\" (UniqueName: \"kubernetes.io/projected/12dcf185-540b-4a55-89e6-cbc3b904a15c-kube-api-access-b7hcs\") pod \"router-default-5444994796-b8244\" (UID: \"12dcf185-540b-4a55-89e6-cbc3b904a15c\") " pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.867663    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.880784    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbj7x\" (UniqueName: \"kubernetes.io/projected/a3f5e5e8-9995-4c7e-b359-f94e232e25fb-kube-api-access-dbj7x\") pod \"cluster-image-registry-operator-dc59b4c8b-k45g7\" (UID: \"a3f5e5e8-9995-4c7e-b359-f94e232e25fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.883110    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26" event={"ID":"bf09d978-94c1-460e-b1ee-29710101ac02","Type":"ContainerStarted","Data":"011a299608a9592def61b680ab7bd2d114d70848729772f486800c662dd0c5ed"}
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.885503    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.38547896 +0000 UTC m=+151.288955963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.885537    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.885926    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.886283    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.386272209 +0000 UTC m=+151.289749202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.887889    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np" event={"ID":"7f5cc342-cace-4433-8fcb-25578dacdb68","Type":"ContainerStarted","Data":"bd90fda2cc9359c86d25d176dff8722f7a8980e0fb51e32aadf9c4349cf78c81"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.895272    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.900301    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.904040    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-bound-sa-token\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.908083    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8878r" event={"ID":"cfe981a6-ed1d-411f-801d-b890b544ee5a","Type":"ContainerStarted","Data":"b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.908326    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8878r" event={"ID":"cfe981a6-ed1d-411f-801d-b890b544ee5a","Type":"ContainerStarted","Data":"d3d480caa0c4757f8d4123d3c8684149691d7c5ccdbe6205c6655e4b7402478a"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.910832    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm" event={"ID":"a2b6bacc-9256-4731-a61c-bbcb743aafe8","Type":"ContainerStarted","Data":"6a7b8914eabd4d38a2e01abafb3a51a67947c79476935cfb56fbdd287117f359"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.910891    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm" event={"ID":"a2b6bacc-9256-4731-a61c-bbcb743aafe8","Type":"ContainerStarted","Data":"d4557e9992e445a4fb6b49c9f4785df8112cd8e1e97d753d8a488f6af122a8ee"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.917730    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mbh2m"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.920901    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.921579    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b95a827d-6910-45c6-8dbf-a45a4cc6578f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-svvpj\" (UID: \"b95a827d-6910-45c6-8dbf-a45a4cc6578f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.923047    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.926161    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.926309    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.933585    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st" event={"ID":"eb655775-5215-4e3c-856b-13a4c678582b","Type":"ContainerStarted","Data":"d1d78f2029ecba2c1b445d5c89957b0d891b4dbc479901669f836650a175ebcc"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.933665    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st" event={"ID":"eb655775-5215-4e3c-856b-13a4c678582b","Type":"ContainerStarted","Data":"f089d930aa2cc19d541c085a58f041c06508bbb78310d66b8edc038bfa1d9bb6"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.933683    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st" event={"ID":"eb655775-5215-4e3c-856b-13a4c678582b","Type":"ContainerStarted","Data":"f9f70b8267d5753cbbae9baab991696050e106bc643213f670ebd1f01be5551d"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.936051    4824 generic.go:334] "Generic (PLEG): container finished" podID="fba4ca2c-4f42-405d-8fc1-60a857973471" containerID="f3652a8ef96c124ca17dbb0071798d70dc1e19c3bc54adbb2b2842b670ecbda2" exitCode=0
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.936156    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7" event={"ID":"fba4ca2c-4f42-405d-8fc1-60a857973471","Type":"ContainerDied","Data":"f3652a8ef96c124ca17dbb0071798d70dc1e19c3bc54adbb2b2842b670ecbda2"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.936222    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7" event={"ID":"fba4ca2c-4f42-405d-8fc1-60a857973471","Type":"ContainerStarted","Data":"13e6ca8e26a07b3f940ba577797c00f706c47b06e8fc127efaf919c7388a4e9f"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.939414    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ssf7\" (UniqueName: \"kubernetes.io/projected/eae9e020-d1a0-467f-b10b-9e022f357fc3-kube-api-access-4ssf7\") pod \"collect-profiles-29484585-7rhkp\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.939647    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wrcr2" event={"ID":"989e489d-6868-4601-82b0-71ee3474a4c1","Type":"ContainerStarted","Data":"49c6343218dd1ccac2218b74d9eebc52bc1b0da36db71b780a6480d8215a3352"}
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.944969    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"
Jan 22 09:45:17 crc kubenswrapper[4824]: W0122 09:45:17.950585    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9adf37cf_0b32_429f_ad06_819052944ee0.slice/crio-20b3d65a7013ea76c29052193d9f3f27ea6410466643c2e129e5df93d402a90b WatchSource:0}: Error finding container 20b3d65a7013ea76c29052193d9f3f27ea6410466643c2e129e5df93d402a90b: Status 404 returned error can't find the container with id 20b3d65a7013ea76c29052193d9f3f27ea6410466643c2e129e5df93d402a90b
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.950927    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f79tb"]
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.953908    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.961693    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skjdt\" (UniqueName: \"kubernetes.io/projected/43bcec80-1438-4a3f-ba1c-48b2055751ac-kube-api-access-skjdt\") pod \"service-ca-operator-777779d784-kb2qp\" (UID: \"43bcec80-1438-4a3f-ba1c-48b2055751ac\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:17 crc kubenswrapper[4824]: W0122 09:45:17.975070    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7525c5ea_34c5_46e7_b49f_0c4a2a06e0fa.slice/crio-ef6a55a122740b7bc92f58769f4211aed5fbec70233313a069ab460ca358a61b WatchSource:0}: Error finding container ef6a55a122740b7bc92f58769f4211aed5fbec70233313a069ab460ca358a61b: Status 404 returned error can't find the container with id ef6a55a122740b7bc92f58769f4211aed5fbec70233313a069ab460ca358a61b
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.983304    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lfjz\" (UniqueName: \"kubernetes.io/projected/cb724a14-a04d-458a-8350-65e8f9d5a69e-kube-api-access-5lfjz\") pod \"machine-config-controller-84d6567774-p4xwj\" (UID: \"cb724a14-a04d-458a-8350-65e8f9d5a69e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:17 crc kubenswrapper[4824]: I0122 09:45:17.986531    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:17 crc kubenswrapper[4824]: E0122 09:45:17.987574    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.487550333 +0000 UTC m=+151.391027326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.003326    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.004469    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h62q\" (UniqueName: \"kubernetes.io/projected/7747087d-a669-4f55-978c-5bce681a7e2e-kube-api-access-9h62q\") pod \"ingress-canary-qdbht\" (UID: \"7747087d-a669-4f55-978c-5bce681a7e2e\") " pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.010444    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.013685    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.022485    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2699m\" (UniqueName: \"kubernetes.io/projected/0e085a8e-5760-4859-bf6b-ac4fac6aa9cf-kube-api-access-2699m\") pod \"multus-admission-controller-857f4d67dd-hpp2x\" (UID: \"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.041493    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46zr8\" (UniqueName: \"kubernetes.io/projected/5ad14705-3767-49da-86e9-544b5259b066-kube-api-access-46zr8\") pod \"olm-operator-6b444d44fb-xl5v9\" (UID: \"5ad14705-3767-49da-86e9-544b5259b066\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.058510    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.059753    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x4jm\" (UniqueName: \"kubernetes.io/projected/3a616855-646b-46d9-a8b4-76130bd53772-kube-api-access-7x4jm\") pod \"machine-config-server-4wqpq\" (UID: \"3a616855-646b-46d9-a8b4-76130bd53772\") " pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.087718    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qdbht"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.088584    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.089078    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.589057085 +0000 UTC m=+151.492534118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.107193    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.111848    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62kg\" (UniqueName: \"kubernetes.io/projected/6803cfcb-a020-4f9c-a668-a2df6abbb407-kube-api-access-m62kg\") pod \"service-ca-9c57cc56f-lprcb\" (UID: \"6803cfcb-a020-4f9c-a668-a2df6abbb407\") " pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.116633    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.118956    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlc4\" (UniqueName: \"kubernetes.io/projected/28ae5f2e-c8dd-4d2c-b318-9677541a6656-kube-api-access-ltlc4\") pod \"kube-storage-version-migrator-operator-b67b599dd-8884b\" (UID: \"28ae5f2e-c8dd-4d2c-b318-9677541a6656\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.122188    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7596399-0f46-4c4a-97ae-379efd65e7af-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf8g8\" (UID: \"a7596399-0f46-4c4a-97ae-379efd65e7af\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.125182    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.141367    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws546\" (UniqueName: \"kubernetes.io/projected/8d492ef8-5e48-4fc7-a56b-1233da10e9f7-kube-api-access-ws546\") pod \"machine-api-operator-5694c8668f-x7psz\" (UID: \"8d492ef8-5e48-4fc7-a56b-1233da10e9f7\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.154182    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.159156    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2b2t\" (UniqueName: \"kubernetes.io/projected/9c2ebdfa-baca-4839-8524-e4161b7cbe89-kube-api-access-b2b2t\") pod \"package-server-manager-789f6589d5-877dz\" (UID: \"9c2ebdfa-baca-4839-8524-e4161b7cbe89\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.178722    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b48m\" (UniqueName: \"kubernetes.io/projected/c8b117b7-e048-4a4a-9592-1d3eb821d309-kube-api-access-5b48m\") pod \"csi-hostpathplugin-hmgmd\" (UID: \"c8b117b7-e048-4a4a-9592-1d3eb821d309\") " pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.189970    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.190546    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.690521845 +0000 UTC m=+151.593998838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.217762    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwkrp\" (UniqueName: \"kubernetes.io/projected/2463610b-364e-4154-83ee-b6c00a6c6da5-kube-api-access-zwkrp\") pod \"marketplace-operator-79b997595-7qjxk\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") " pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.231426    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.238550    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.239881    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52kr2\" (UniqueName: \"kubernetes.io/projected/faa5f086-812d-4c56-9491-fd7147a6706a-kube-api-access-52kr2\") pod \"catalog-operator-68c6474976-rgv6r\" (UID: \"faa5f086-812d-4c56-9491-fd7147a6706a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.245244    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.252103    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvtfl\" (UniqueName: \"kubernetes.io/projected/97038c7c-c8e0-451d-86a5-93e5f36d5e33-kube-api-access-tvtfl\") pod \"packageserver-d55dfcdfc-wgvg2\" (UID: \"97038c7c-c8e0-451d-86a5-93e5f36d5e33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.263985    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"
Jan 22 09:45:18 crc kubenswrapper[4824]: W0122 09:45:18.268052    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1215925_7f63_449c_8aa1_879ac01ed726.slice/crio-51df50ffb85c316b0d714b6ff62a3d3cafdc89e435ea35955b11293042b2e4c6 WatchSource:0}: Error finding container 51df50ffb85c316b0d714b6ff62a3d3cafdc89e435ea35955b11293042b2e4c6: Status 404 returned error can't find the container with id 51df50ffb85c316b0d714b6ff62a3d3cafdc89e435ea35955b11293042b2e4c6
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.268437    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.273601    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.279541    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l49bn\" (UniqueName: \"kubernetes.io/projected/4809673b-3d53-4bd5-8739-9be6a2443e60-kube-api-access-l49bn\") pod \"downloads-7954f5f757-rmsj9\" (UID: \"4809673b-3d53-4bd5-8739-9be6a2443e60\") " pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.280654    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.289292    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.294443    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz98s\" (UniqueName: \"kubernetes.io/projected/29832635-7ac3-4472-afcd-3f06b7a2fe91-kube-api-access-qz98s\") pod \"dns-default-zcblz\" (UID: \"29832635-7ac3-4472-afcd-3f06b7a2fe91\") " pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.294732    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.295575    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.296011    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.795993505 +0000 UTC m=+151.699470498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.321197    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-lprcb"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.325165    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4wqpq"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.374114    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.378383    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.388394    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.396728    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.397898    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:18.897860541 +0000 UTC m=+151.801337544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.475550    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.500107    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.500546    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.000528466 +0000 UTC m=+151.904005459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.553822    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.577047    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.601041    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.602432    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.102411142 +0000 UTC m=+152.005888135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: W0122 09:45:18.649911    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78d695a9_430a_4540_8a70_5c2bcfc40b34.slice/crio-f733e0775cf67dc06804f2a027c1940077b00daadd8a5a8c1c922457a63de6b0 WatchSource:0}: Error finding container f733e0775cf67dc06804f2a027c1940077b00daadd8a5a8c1c922457a63de6b0: Status 404 returned error can't find the container with id f733e0775cf67dc06804f2a027c1940077b00daadd8a5a8c1c922457a63de6b0
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.704202    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.704589    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.204571209 +0000 UTC m=+152.108048202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.806903    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.807849    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.808299    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.308276933 +0000 UTC m=+152.211753926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.872919    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l9tjj"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.909796    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:18 crc kubenswrapper[4824]: E0122 09:45:18.910228    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.41019317 +0000 UTC m=+152.313670163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.947733    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz"]
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.963439    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf" event={"ID":"28370e19-78ca-49e9-916f-aaa80793d644","Type":"ContainerStarted","Data":"c0c4a67e09c9e8f2db809242089ba3a26ea56316b772c36c2d9cf5978cadc0a0"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.965370    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" event={"ID":"78d695a9-430a-4540-8a70-5c2bcfc40b34","Type":"ContainerStarted","Data":"f733e0775cf67dc06804f2a027c1940077b00daadd8a5a8c1c922457a63de6b0"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.970300    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj" event={"ID":"4533ce52-cfcf-4df1-9f6f-e8d6febd7719","Type":"ContainerStarted","Data":"cb4985cb18e4735292e9d804350ff92705795539fa001f3c03880c9ddb3dc164"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.972771    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wrcr2" event={"ID":"989e489d-6868-4601-82b0-71ee3474a4c1","Type":"ContainerStarted","Data":"2096010d8d75091f7450cf683ffcf3ba520d9babce61cb7205328368e62167da"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.972969    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.974089    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj" event={"ID":"b5902322-e067-4ed7-967d-f71553df3532","Type":"ContainerStarted","Data":"9ba41c9d40664aee4e7e85ca9ad37debfea3b89f78f0063e03bf496d4cb14c86"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.977117    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f79tb" event={"ID":"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa","Type":"ContainerStarted","Data":"ef6a55a122740b7bc92f58769f4211aed5fbec70233313a069ab460ca358a61b"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.978060    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk" event={"ID":"602843d5-8e62-4834-a7ae-af01d9f14644","Type":"ContainerStarted","Data":"9c31c48d6646ef9ac82884077318a25e4e8493dd038d2ad8d96d49ba466ed10d"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.978814    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv" event={"ID":"697490dd-5ba8-4719-b8a3-87a9b9ed08b3","Type":"ContainerStarted","Data":"0acecaf27b05ea7706845c0fae7efcc9269d6579f576cd2d5c7fb04fb5430c44"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.979904    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m" event={"ID":"9adf37cf-0b32-429f-ad06-819052944ee0","Type":"ContainerStarted","Data":"20b3d65a7013ea76c29052193d9f3f27ea6410466643c2e129e5df93d402a90b"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.980630    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x" event={"ID":"a1215925-7f63-449c-8aa1-879ac01ed726","Type":"ContainerStarted","Data":"51df50ffb85c316b0d714b6ff62a3d3cafdc89e435ea35955b11293042b2e4c6"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.981668    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26" event={"ID":"bf09d978-94c1-460e-b1ee-29710101ac02","Type":"ContainerStarted","Data":"85efd8af970fb76d1905f5248095046e4169c6a99b156420dea34b9c18097f72"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.986574    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np" event={"ID":"7f5cc342-cace-4433-8fcb-25578dacdb68","Type":"ContainerStarted","Data":"c10d48f23a58ec6a7341fb13b0deabf111b87b1de0d8ad5b930fee8eb3c8bbf5"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.990498    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-b8244" event={"ID":"12dcf185-540b-4a55-89e6-cbc3b904a15c","Type":"ContainerStarted","Data":"89e3f2d582682bb9e346f9dc9e83deae637ba1d2d5edf2d6137ffd9f8fa230bb"}
Jan 22 09:45:18 crc kubenswrapper[4824]: I0122 09:45:18.992713    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4wqpq" event={"ID":"3a616855-646b-46d9-a8b4-76130bd53772","Type":"ContainerStarted","Data":"498f7ff72b43975667bc4efccf10439ef95e25536b1d62113c867dcaef1cde83"}
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.011721    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.011963    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.511940291 +0000 UTC m=+152.415417284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.012050    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.012428    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.512420819 +0000 UTC m=+152.415897812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.124598    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.126114    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.626091746 +0000 UTC m=+152.529568739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.228174    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.228944    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.728928777 +0000 UTC m=+152.632405770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.312622    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6vpgm" podStartSLOduration=127.312596933 podStartE2EDuration="2m7.312596933s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:19.311415939 +0000 UTC m=+152.214892942" watchObservedRunningTime="2026-01-22 09:45:19.312596933 +0000 UTC m=+152.216073926"
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.329856    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.330338    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.830312295 +0000 UTC m=+152.733789298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.337404    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8878r" podStartSLOduration=127.337383969 podStartE2EDuration="2m7.337383969s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:19.335119394 +0000 UTC m=+152.238596387" watchObservedRunningTime="2026-01-22 09:45:19.337383969 +0000 UTC m=+152.240860962"
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.380598    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.382233    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wrcr2" podStartSLOduration=127.382213044 podStartE2EDuration="2m7.382213044s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:19.377591871 +0000 UTC m=+152.281068864" watchObservedRunningTime="2026-01-22 09:45:19.382213044 +0000 UTC m=+152.285690037"
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.384337    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-hpp2x"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.429610    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.434189    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.434603    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:19.93458844 +0000 UTC m=+152.838065433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.449131    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.535082    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.535329    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.035285762 +0000 UTC m=+152.938762755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.535599    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.536039    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.03603019 +0000 UTC m=+152.939507183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.625474    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6m2np" podStartSLOduration=127.625449151 podStartE2EDuration="2m7.625449151s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:19.624006077 +0000 UTC m=+152.527483090" watchObservedRunningTime="2026-01-22 09:45:19.625449151 +0000 UTC m=+152.528926144"
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.636908    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.637397    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.137329124 +0000 UTC m=+153.040806117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: W0122 09:45:19.652761    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e085a8e_5760_4859_bf6b_ac4fac6aa9cf.slice/crio-571265668d65aef958b091c698275c0e4f158b6253d06ab0d7c2621a236d5015 WatchSource:0}: Error finding container 571265668d65aef958b091c698275c0e4f158b6253d06ab0d7c2621a236d5015: Status 404 returned error can't find the container with id 571265668d65aef958b091c698275c0e4f158b6253d06ab0d7c2621a236d5015
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.724035    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xk8qh"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.738711    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.739141    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.239121107 +0000 UTC m=+153.142598100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.755288    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qdbht"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.756765    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x7psz"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.785980    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zkj26" podStartSLOduration=127.785946116 podStartE2EDuration="2m7.785946116s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:19.779834408 +0000 UTC m=+152.683311411" watchObservedRunningTime="2026-01-22 09:45:19.785946116 +0000 UTC m=+152.689423119"
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.840900    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.841252    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.341228792 +0000 UTC m=+153.244705785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.872101    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj"]
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.910858    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wrcr2"
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.911583    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"]
Jan 22 09:45:19 crc kubenswrapper[4824]: W0122 09:45:19.919983    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb724a14_a04d_458a_8350_65e8f9d5a69e.slice/crio-b0052801dc417a7b4c016e8c220ffbe192144c9dcf474bc7c99201526b93828f WatchSource:0}: Error finding container b0052801dc417a7b4c016e8c220ffbe192144c9dcf474bc7c99201526b93828f: Status 404 returned error can't find the container with id b0052801dc417a7b4c016e8c220ffbe192144c9dcf474bc7c99201526b93828f
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.942554    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:19 crc kubenswrapper[4824]: E0122 09:45:19.942947    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.442931991 +0000 UTC m=+153.346408984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:19 crc kubenswrapper[4824]: I0122 09:45:19.967314    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zcblz"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:19.982667    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lw2st" podStartSLOduration=127.982645925 podStartE2EDuration="2m7.982645925s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:19.981930208 +0000 UTC m=+152.885407201" watchObservedRunningTime="2026-01-22 09:45:19.982645925 +0000 UTC m=+152.886122918"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.024246    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.026295    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hmgmd"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.042156    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.043846    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.044204    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.544182153 +0000 UTC m=+153.447659136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.047405    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rmsj9"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.059448    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-822xl"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.111543    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz" event={"ID":"526c6812-a825-4d53-90fb-fa01195fa33a","Type":"ContainerStarted","Data":"d247e0bf4f5cc5f7ac9999d8d6c6c50ae59ad846310e8a084554cab4f5fb9ee1"}
Jan 22 09:45:20 crc kubenswrapper[4824]: W0122 09:45:20.125777    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc660b59_eb41_4c18_84a4_ebd782752570.slice/crio-fd63951e89ff43f969261b165cad15530dae87833644b179376372eecf9c8399 WatchSource:0}: Error finding container fd63951e89ff43f969261b165cad15530dae87833644b179376372eecf9c8399: Status 404 returned error can't find the container with id fd63951e89ff43f969261b165cad15530dae87833644b179376372eecf9c8399
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.136184    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.136850    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj" event={"ID":"37205799-5637-4c12-8215-16f21fe22c72","Type":"ContainerStarted","Data":"b1abb232b942f25e62aab94ad52f5ac5632b3503546731b5713168b8c605a0a4"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.151980    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.152811    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.652795471 +0000 UTC m=+153.556272464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.160458    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7qjxk"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.164847    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.176551    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b" event={"ID":"28ae5f2e-c8dd-4d2c-b318-9677541a6656","Type":"ContainerStarted","Data":"70a3a58d4c82380d6e0ad640720e83a07780c41a81c54c0febf209b9129b73b8"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.180496    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qdbht" event={"ID":"7747087d-a669-4f55-978c-5bce681a7e2e","Type":"ContainerStarted","Data":"e0785e10ee76800e2adf9432379cac676389a59000cef3e38b4e33d19b52b149"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.203527    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7" event={"ID":"a3f5e5e8-9995-4c7e-b359-f94e232e25fb","Type":"ContainerStarted","Data":"3228565cf8d79c2ff3f35fac8a4948f10c3c65f4d03dc9fc56bee083cc783826"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.216586    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp" event={"ID":"43bcec80-1438-4a3f-ba1c-48b2055751ac","Type":"ContainerStarted","Data":"9af76a39ba4846c97b383e25c4f4337732212c8bec1ffdb6bf3a9a93d8d9e676"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.224633    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"]
Jan 22 09:45:20 crc kubenswrapper[4824]: W0122 09:45:20.236422    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4809673b_3d53_4bd5_8739_9be6a2443e60.slice/crio-ad0b4ba525b623a0836672be2b427e29186c225f65f614a079dc143964f52b56 WatchSource:0}: Error finding container ad0b4ba525b623a0836672be2b427e29186c225f65f614a079dc143964f52b56: Status 404 returned error can't find the container with id ad0b4ba525b623a0836672be2b427e29186c225f65f614a079dc143964f52b56
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.237350    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh" event={"ID":"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f","Type":"ContainerStarted","Data":"69b484aadce347cd795e85bd7fc30073ba4f8c1895a209449b28438e12c54491"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.254844    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lprcb"]
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.257279    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.258424    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.758329463 +0000 UTC m=+153.661806466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.263910    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7" event={"ID":"fba4ca2c-4f42-405d-8fc1-60a857973471","Type":"ContainerStarted","Data":"3f08687e87ce39e463f7972e3b92b157d22107f1d0cf6e551623eff68c8dda77"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.311979    4824 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5n9hc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body=
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.312049    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" podUID="78d695a9-430a-4540-8a70-5c2bcfc40b34" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.313806    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.360032    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.360299    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.360340    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.362485    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.862462883 +0000 UTC m=+153.765939886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.368511    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk" event={"ID":"602843d5-8e62-4834-a7ae-af01d9f14644","Type":"ContainerStarted","Data":"192d96a534a2d5b0615f0b91b8a7d8bbbec5228b6cc11ce49aeb27abd2825ea6"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.368647    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.382873    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.387174    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz" event={"ID":"8d492ef8-5e48-4fc7-a56b-1233da10e9f7","Type":"ContainerStarted","Data":"7847abcf78844ee321503990aec6405e92b13ff91ca47a17865d658b706631f7"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.405637    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj" event={"ID":"cb724a14-a04d-458a-8350-65e8f9d5a69e","Type":"ContainerStarted","Data":"b0052801dc417a7b4c016e8c220ffbe192144c9dcf474bc7c99201526b93828f"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.444275    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.463246    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.463673    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.463860    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.468573    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:20.968527626 +0000 UTC m=+153.872004769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.472888    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.477039    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.481793    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9" event={"ID":"5ad14705-3767-49da-86e9-544b5259b066","Type":"ContainerStarted","Data":"283ccf6a656a7334d3e22835750ebfa386342ea6b2b50b1a054d182b898d0b9d"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.507584    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x" event={"ID":"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf","Type":"ContainerStarted","Data":"571265668d65aef958b091c698275c0e4f158b6253d06ab0d7c2621a236d5015"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.521662    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj" event={"ID":"b5902322-e067-4ed7-967d-f71553df3532","Type":"ContainerStarted","Data":"83a29f695280e413dbdf1ec14910269d7005579f831cd6573426a5e3b86dc8c2"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.543749    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp" event={"ID":"eae9e020-d1a0-467f-b10b-9e022f357fc3","Type":"ContainerStarted","Data":"5c37d5229908deafd9dfb7877a5bd261fd62b4b8d2e6044d8974d7ed605e2c1d"}
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.565250    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.566070    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.066033318 +0000 UTC m=+153.969510311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.675945    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.677456    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.177347147 +0000 UTC m=+154.080824140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.722621    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp" podStartSLOduration=20.722595927 podStartE2EDuration="20.722595927s" podCreationTimestamp="2026-01-22 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:20.720839892 +0000 UTC m=+153.624316905" watchObservedRunningTime="2026-01-22 09:45:20.722595927 +0000 UTC m=+153.626072930"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.726890    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.749600    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.777956    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.778528    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.278508816 +0000 UTC m=+154.181985809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.800393    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" podStartSLOduration=128.800345782 podStartE2EDuration="2m8.800345782s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:20.798783363 +0000 UTC m=+153.702260356" watchObservedRunningTime="2026-01-22 09:45:20.800345782 +0000 UTC m=+153.703822775"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.847343    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zf5hj" podStartSLOduration=128.847322107 podStartE2EDuration="2m8.847322107s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:20.844904826 +0000 UTC m=+153.748381829" watchObservedRunningTime="2026-01-22 09:45:20.847322107 +0000 UTC m=+153.750799110"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.879632    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.880083    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.38005996 +0000 UTC m=+154.283536963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.943706    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7" podStartSLOduration=128.943653435 podStartE2EDuration="2m8.943653435s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:20.936872012 +0000 UTC m=+153.840349025" watchObservedRunningTime="2026-01-22 09:45:20.943653435 +0000 UTC m=+153.847130428"
Jan 22 09:45:20 crc kubenswrapper[4824]: I0122 09:45:20.985117    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:20 crc kubenswrapper[4824]: E0122 09:45:20.985527    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.485511439 +0000 UTC m=+154.388988432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.088437    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.089089    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.589069028 +0000 UTC m=+154.492546021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.190056    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.190525    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.690509487 +0000 UTC m=+154.593986480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.300922    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.301617    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.801596216 +0000 UTC m=+154.705073209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.407176    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.407660    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:21.907643008 +0000 UTC m=+154.811120001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.507855    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.508255    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.008222335 +0000 UTC m=+154.911699328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.508740    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.509035    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.009027145 +0000 UTC m=+154.912504138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.579145    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj" event={"ID":"cb724a14-a04d-458a-8350-65e8f9d5a69e","Type":"ContainerStarted","Data":"56f6ec18e12146aa23ab920b8f0ab38fc0a72e04bd5de442d1a4a42ffc201201"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.605818    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp" event={"ID":"43bcec80-1438-4a3f-ba1c-48b2055751ac","Type":"ContainerStarted","Data":"7fe7e73813ee9801b449c50284fcce41f3e377618478ded51a6ea1d7e281a851"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.610524    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.610884    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.110823298 +0000 UTC m=+155.014300291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.610966    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.611260    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.111246834 +0000 UTC m=+155.014723827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.612759    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd" event={"ID":"c8b117b7-e048-4a4a-9592-1d3eb821d309","Type":"ContainerStarted","Data":"9e1873a7bd76c484b2be8fe167fc4a6ac173a8b720ffc5c81a5986f2a873e3b9"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.613811    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f79tb" event={"ID":"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa","Type":"ContainerStarted","Data":"467b72c067b29a2a56bc7d4e10167df85996d81ba9d06014a1229fab42323dc1"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.615125    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b" event={"ID":"28ae5f2e-c8dd-4d2c-b318-9677541a6656","Type":"ContainerStarted","Data":"2f294b116e6176af269870c757adc04d744781dbfd94dc2aa209e63d314c8b3f"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.640786    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kb2qp" podStartSLOduration=129.640762326 podStartE2EDuration="2m9.640762326s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.636278159 +0000 UTC m=+154.539755152" watchObservedRunningTime="2026-01-22 09:45:21.640762326 +0000 UTC m=+154.544239319"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.687629    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh" event={"ID":"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f","Type":"ContainerStarted","Data":"8635b27ca4ff819e02855846c1c473131753df1e9c83d5a4b11d210e39b2e430"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.688891    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.711526    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.712711    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.212691084 +0000 UTC m=+155.116168077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.714005    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qdbht" event={"ID":"7747087d-a669-4f55-978c-5bce681a7e2e","Type":"ContainerStarted","Data":"a1a294d9fb3fd15f0ba301dda4b5da157b5bdd6ece970b93ac29578ad42bd6d4"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.719612    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.730639    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8884b" podStartSLOduration=129.730621783 podStartE2EDuration="2m9.730621783s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.685535829 +0000 UTC m=+154.589012822" watchObservedRunningTime="2026-01-22 09:45:21.730621783 +0000 UTC m=+154.634098776"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.731778    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh" podStartSLOduration=129.731773576 podStartE2EDuration="2m9.731773576s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.728817186 +0000 UTC m=+154.632294179" watchObservedRunningTime="2026-01-22 09:45:21.731773576 +0000 UTC m=+154.635250569"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.749794    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r" event={"ID":"faa5f086-812d-4c56-9491-fd7147a6706a","Type":"ContainerStarted","Data":"94aa4e67baf4591a3b9d0fe41133acc024c7fe6c76a2988ce76fb36134560259"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.750073    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r" event={"ID":"faa5f086-812d-4c56-9491-fd7147a6706a","Type":"ContainerStarted","Data":"278b1b36526445ca611a7f19c7711781d71de7c820763ef4c2f7a61524bee8fd"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.751256    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.754190    4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rgv6r container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body=
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.754296    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r" podUID="faa5f086-812d-4c56-9491-fd7147a6706a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.759352    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qdbht" podStartSLOduration=6.759323366 podStartE2EDuration="6.759323366s" podCreationTimestamp="2026-01-22 09:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.757219397 +0000 UTC m=+154.660696390" watchObservedRunningTime="2026-01-22 09:45:21.759323366 +0000 UTC m=+154.662800369"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.796613    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj" event={"ID":"37205799-5637-4c12-8215-16f21fe22c72","Type":"ContainerStarted","Data":"b51e1d2c6acec12ca2bac3fe2b4bb41931c92e78ffb79308ffe50d918f79f958"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.798800    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj" event={"ID":"b95a827d-6910-45c6-8dbf-a45a4cc6578f","Type":"ContainerStarted","Data":"d72f331e85f6735e7db2b4a1a17500a34ef39f20af4f0c5e3fcd4c329375b9c6"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.798821    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj" event={"ID":"b95a827d-6910-45c6-8dbf-a45a4cc6578f","Type":"ContainerStarted","Data":"6874ee647103b55e4fcb2a1a86db4ea89fee622aac3c5a66f46603cb275a0ff9"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.801851    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x" event={"ID":"a1215925-7f63-449c-8aa1-879ac01ed726","Type":"ContainerStarted","Data":"b2a443c27adc925e2cdf547f2508ea2e253fe2e5993f9e7c0953e8701e51cefc"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.813440    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.816647    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.316623346 +0000 UTC m=+155.220100399 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.825205    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r" podStartSLOduration=129.825167315 podStartE2EDuration="2m9.825167315s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.824543032 +0000 UTC m=+154.728020025" watchObservedRunningTime="2026-01-22 09:45:21.825167315 +0000 UTC m=+154.728644308"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.828171    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rmsj9" event={"ID":"4809673b-3d53-4bd5-8739-9be6a2443e60","Type":"ContainerStarted","Data":"ad0b4ba525b623a0836672be2b427e29186c225f65f614a079dc143964f52b56"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.849710    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-svvpj" podStartSLOduration=129.849687331 podStartE2EDuration="2m9.849687331s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.848299479 +0000 UTC m=+154.751776482" watchObservedRunningTime="2026-01-22 09:45:21.849687331 +0000 UTC m=+154.753164324"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.864722    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz" event={"ID":"9c2ebdfa-baca-4839-8524-e4161b7cbe89","Type":"ContainerStarted","Data":"4e06ae0704d371661d106fa85e713104d780ccc2120b30ce64a75fe9c6ddfb2f"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.864779    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz" event={"ID":"9c2ebdfa-baca-4839-8524-e4161b7cbe89","Type":"ContainerStarted","Data":"737fae683e076a9a2fc9a575a7be6af07f7e18c70120ccc8d31b43ebe3680c3e"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.907854    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-l9tjj" podStartSLOduration=129.907831473 podStartE2EDuration="2m9.907831473s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:21.907818343 +0000 UTC m=+154.811295356" watchObservedRunningTime="2026-01-22 09:45:21.907831473 +0000 UTC m=+154.811308466"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.919600    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-b8244" event={"ID":"12dcf185-540b-4a55-89e6-cbc3b904a15c","Type":"ContainerStarted","Data":"91f5f15fe97cdd66bc7a3ad91130caad855b64c2b558e759e888802dc3836c9e"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.922981    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:21 crc kubenswrapper[4824]: E0122 09:45:21.924758    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.424717524 +0000 UTC m=+155.328194517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.941548    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4wqpq" event={"ID":"3a616855-646b-46d9-a8b4-76130bd53772","Type":"ContainerStarted","Data":"bfaa87a46961a22dd51d1167bea9647ca660a71c867bd2431a98143b6b6cd14e"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.960744    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" event={"ID":"2463610b-364e-4154-83ee-b6c00a6c6da5","Type":"ContainerStarted","Data":"dfcdc054935a47b5bd5618fc0889b41ccbfa317645c7dad101da36f76d555002"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.961908    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.963578    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9" event={"ID":"5ad14705-3767-49da-86e9-544b5259b066","Type":"ContainerStarted","Data":"8b9bdf9585bf321a74e098c2a402f1a4190a1ad05c619c77c66cfa522037fe1c"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.964244    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.965272    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7" event={"ID":"a3f5e5e8-9995-4c7e-b359-f94e232e25fb","Type":"ContainerStarted","Data":"0f121916a5cb7230ae2247e06f36ada9e7756bc5f589e66933aee6b77e7f3cc3"}
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.965889    4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7qjxk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body=
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.965923    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.983513    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.983904    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:21 crc kubenswrapper[4824]: I0122 09:45:21.987664    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x" event={"ID":"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf","Type":"ContainerStarted","Data":"27b8e9eabe476eab92be0c72aae7e5b08aeb9b2a4412d1e8c6a6761a4c527abc"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.048784    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.057031    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.060459    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-lprcb" event={"ID":"6803cfcb-a020-4f9c-a668-a2df6abbb407","Type":"ContainerStarted","Data":"8bc574a300d1f584ec15286bcc251fa023b0aff506d12880169e0518942a5050"}
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.061207    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.561189442 +0000 UTC m=+155.464666435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.062085    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.072620    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" podStartSLOduration=130.072582988 podStartE2EDuration="2m10.072582988s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.059402716 +0000 UTC m=+154.962879709" watchObservedRunningTime="2026-01-22 09:45:22.072582988 +0000 UTC m=+154.976060001"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.072995    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-b8244" podStartSLOduration=130.072988983 podStartE2EDuration="2m10.072988983s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.006853383 +0000 UTC m=+154.910330386" watchObservedRunningTime="2026-01-22 09:45:22.072988983 +0000 UTC m=+154.976465976"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.119028    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2" event={"ID":"97038c7c-c8e0-451d-86a5-93e5f36d5e33","Type":"ContainerStarted","Data":"e50852384f484d385049e7257499c078f22999ebddc4112478f4443b3b5b6557"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.120308    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.148204    4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wgvg2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body=
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.148265    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2" podUID="97038c7c-c8e0-451d-86a5-93e5f36d5e33" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.156483    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv" event={"ID":"697490dd-5ba8-4719-b8a3-87a9b9ed08b3","Type":"ContainerStarted","Data":"094c6d43964c09c78082fe6bb327768411666f70af4f53730e95efdecdcdbaff"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.158138    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.160422    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.660400049 +0000 UTC m=+155.563877042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.188866    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xl5v9" podStartSLOduration=130.188833661 podStartE2EDuration="2m10.188833661s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.124678394 +0000 UTC m=+155.028155387" watchObservedRunningTime="2026-01-22 09:45:22.188833661 +0000 UTC m=+155.092310654"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.190925    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8" event={"ID":"a7596399-0f46-4c4a-97ae-379efd65e7af","Type":"ContainerStarted","Data":"a841edc316d6663dfe6fc6edfd36bb1647f24fae60c2da591050125a58246c1a"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.210300    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4wqpq" podStartSLOduration=7.210276152 podStartE2EDuration="7.210276152s" podCreationTimestamp="2026-01-22 09:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.167597088 +0000 UTC m=+155.071074081" watchObservedRunningTime="2026-01-22 09:45:22.210276152 +0000 UTC m=+155.113753145"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.214552    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" event={"ID":"78d695a9-430a-4540-8a70-5c2bcfc40b34","Type":"ContainerStarted","Data":"25bbcfb7b006dc80251509eba288add7fa18779d72388522ce04ae6d15473b35"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.217704    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k45g7" podStartSLOduration=130.217684999 podStartE2EDuration="2m10.217684999s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.213286014 +0000 UTC m=+155.116762997" watchObservedRunningTime="2026-01-22 09:45:22.217684999 +0000 UTC m=+155.121161992"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.223710    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.271502    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.276770    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj" event={"ID":"4533ce52-cfcf-4df1-9f6f-e8d6febd7719","Type":"ContainerStarted","Data":"a4769c0538c9ed0b4e2959688b722c869e22ff9a1df5a8a7b5e1d1f32b8f2e8f"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.276839    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj" event={"ID":"4533ce52-cfcf-4df1-9f6f-e8d6febd7719","Type":"ContainerStarted","Data":"e95e6fbdb7253d7c5efc3dacebb8425af4b1d89db2dd65540c0332b12cd5c1f6"}
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.279573    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.779534009 +0000 UTC m=+155.683011002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.303515    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz" event={"ID":"526c6812-a825-4d53-90fb-fa01195fa33a","Type":"ContainerStarted","Data":"4683b7148b60f92934ba5f3b8a195bb0248c0d874fe43d6354f4ccc220414ecd"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.308036    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8" podStartSLOduration=130.308002163 podStartE2EDuration="2m10.308002163s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.251868976 +0000 UTC m=+155.155345969" watchObservedRunningTime="2026-01-22 09:45:22.308002163 +0000 UTC m=+155.211479176"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.365014    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf" event={"ID":"28370e19-78ca-49e9-916f-aaa80793d644","Type":"ContainerStarted","Data":"144e33bd4ee79786be4186334fc6d542d366aac072dd63ebdb45100e79a0e0c3"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.365411    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf" event={"ID":"28370e19-78ca-49e9-916f-aaa80793d644","Type":"ContainerStarted","Data":"97e26661b938b0abb81a6bc1d7167161aaa84eb3794bdd09d8fdafa3a99da3da"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.381715    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv" podStartSLOduration=130.381690656 podStartE2EDuration="2m10.381690656s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.31034456 +0000 UTC m=+155.213821573" watchObservedRunningTime="2026-01-22 09:45:22.381690656 +0000 UTC m=+155.285167649"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.384239    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.385218    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.885198017 +0000 UTC m=+155.788675010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.387331    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp" event={"ID":"eae9e020-d1a0-467f-b10b-9e022f357fc3","Type":"ContainerStarted","Data":"23f6131210ede6ea7ba593143c84c75a67839933dcb9683b76482ec317d75e0d"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.432176    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" event={"ID":"fc660b59-eb41-4c18-84a4-ebd782752570","Type":"ContainerStarted","Data":"fd63951e89ff43f969261b165cad15530dae87833644b179376372eecf9c8399"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.434336    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.452006    4824 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-822xl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" start-of-body=
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.452076    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.480004    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zcblz" event={"ID":"29832635-7ac3-4472-afcd-3f06b7a2fe91","Type":"ContainerStarted","Data":"fb9f2eab3f05afa4160c473f1888357ecf8c2b22ffa34b49fb47201799f0aeaa"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.493261    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.494862    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:22.994838072 +0000 UTC m=+155.898315065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.499017    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"32feaa26a507d920ba0d9beea702198279ef981b1ea5b1abbcc7d077a8579686"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.572007    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2" podStartSLOduration=130.571986765 podStartE2EDuration="2m10.571986765s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.559419985 +0000 UTC m=+155.462896978" watchObservedRunningTime="2026-01-22 09:45:22.571986765 +0000 UTC m=+155.475463758"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.573260    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz" event={"ID":"8d492ef8-5e48-4fc7-a56b-1233da10e9f7","Type":"ContainerStarted","Data":"ebb69b8dda2f73911fda790fcf740d37fec59b7abb0254face6f8ee794f13fab"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.582059    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-lprcb" podStartSLOduration=130.58203413 podStartE2EDuration="2m10.58203413s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.579864609 +0000 UTC m=+155.483341602" watchObservedRunningTime="2026-01-22 09:45:22.58203413 +0000 UTC m=+155.485511133"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.596983    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.599898    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.099874786 +0000 UTC m=+156.003351779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.619110    4824 generic.go:334] "Generic (PLEG): container finished" podID="602843d5-8e62-4834-a7ae-af01d9f14644" containerID="192d96a534a2d5b0615f0b91b8a7d8bbbec5228b6cc11ce49aeb27abd2825ea6" exitCode=0
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.619479    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk" event={"ID":"602843d5-8e62-4834-a7ae-af01d9f14644","Type":"ContainerDied","Data":"192d96a534a2d5b0615f0b91b8a7d8bbbec5228b6cc11ce49aeb27abd2825ea6"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.619551    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.672815    4824 generic.go:334] "Generic (PLEG): container finished" podID="9adf37cf-0b32-429f-ad06-819052944ee0" containerID="1bee609ab2649ce7762f74a0e8b4c13bc123c2cc827136d67252762fede0a595" exitCode=0
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.673021    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m" event={"ID":"9adf37cf-0b32-429f-ad06-819052944ee0","Type":"ContainerDied","Data":"1bee609ab2649ce7762f74a0e8b4c13bc123c2cc827136d67252762fede0a595"}
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.691755    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlzt7"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.700085    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.715349    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.201331777 +0000 UTC m=+156.104808770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.757234    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-97mmf" podStartSLOduration=130.757212204 podStartE2EDuration="2m10.757212204s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.757017547 +0000 UTC m=+155.660494560" watchObservedRunningTime="2026-01-22 09:45:22.757212204 +0000 UTC m=+155.660689197"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.803906    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.806309    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.306279537 +0000 UTC m=+156.209756530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.871016    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.874486    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:22 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:22 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:22 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.874556    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.906562    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tkrmz" podStartSLOduration=130.906540153 podStartE2EDuration="2m10.906540153s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.82268517 +0000 UTC m=+155.726162163" watchObservedRunningTime="2026-01-22 09:45:22.906540153 +0000 UTC m=+155.810017146"
Jan 22 09:45:22 crc kubenswrapper[4824]: I0122 09:45:22.908097    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:22 crc kubenswrapper[4824]: E0122 09:45:22.908461    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.408445874 +0000 UTC m=+156.311922867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.002202    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-zg6fj" podStartSLOduration=131.002175185 podStartE2EDuration="2m11.002175185s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:22.966367738 +0000 UTC m=+155.869844731" watchObservedRunningTime="2026-01-22 09:45:23.002175185 +0000 UTC m=+155.905652178"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.009784    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.010032    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.509999218 +0000 UTC m=+156.413476211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.010164    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.010608    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.51060013 +0000 UTC m=+156.414077123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.089623    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" podStartSLOduration=131.089602681 podStartE2EDuration="2m11.089602681s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.0884971 +0000 UTC m=+155.991974093" watchObservedRunningTime="2026-01-22 09:45:23.089602681 +0000 UTC m=+155.993079674"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.090832    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz" podStartSLOduration=131.090827667 podStartE2EDuration="2m11.090827667s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.018672192 +0000 UTC m=+155.922149185" watchObservedRunningTime="2026-01-22 09:45:23.090827667 +0000 UTC m=+155.994304660"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.134275    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.134988    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.634962496 +0000 UTC m=+156.538439489 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.135803    4824 csr.go:261] certificate signing request csr-tk52h is approved, waiting to be issued
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.174018    4824 csr.go:257] certificate signing request csr-tk52h is issued
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.233017    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk" podStartSLOduration=131.232993608 podStartE2EDuration="2m11.232993608s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.232889224 +0000 UTC m=+156.136366217" watchObservedRunningTime="2026-01-22 09:45:23.232993608 +0000 UTC m=+156.136470601"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.236207    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.236737    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.736719167 +0000 UTC m=+156.640196160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.337261    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.337636    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.837617417 +0000 UTC m=+156.741094410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.438543    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.438924    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:23.938909761 +0000 UTC m=+156.842386754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.539292    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.539435    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.039400295 +0000 UTC m=+156.942877288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.539708    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.540192    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.040167253 +0000 UTC m=+156.943644306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.641340    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.641557    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.14151822 +0000 UTC m=+157.044995223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.641642    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.642123    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.142106141 +0000 UTC m=+157.045583144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.693588    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zcblz" event={"ID":"29832635-7ac3-4472-afcd-3f06b7a2fe91","Type":"ContainerStarted","Data":"4c735b0c93621cbc660c856186fda524bf0086cab1e7571f138af4da63cd83ed"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.693916    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zcblz" event={"ID":"29832635-7ac3-4472-afcd-3f06b7a2fe91","Type":"ContainerStarted","Data":"186fe6cb411b21526461c0c101033e14bf9200789e89d51fd860801e54d95c8d"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.693933    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.701373    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd" event={"ID":"c8b117b7-e048-4a4a-9592-1d3eb821d309","Type":"ContainerStarted","Data":"bed52a89af40b638ab0c03ab20e34381a0d1931909d4774e0675bd2967fe2d46"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.701439    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd" event={"ID":"c8b117b7-e048-4a4a-9592-1d3eb821d309","Type":"ContainerStarted","Data":"1e132364d880a5d9ee9c29f03a26d0bbb20dd3dd1fcd9be5955fd792d3d65248"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.703753    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2" event={"ID":"97038c7c-c8e0-451d-86a5-93e5f36d5e33","Type":"ContainerStarted","Data":"05b0eb0d6633965ca8034d782f1011665f0444a6e31e4a0db6427eb2d06c5845"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.706971    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m" event={"ID":"9adf37cf-0b32-429f-ad06-819052944ee0","Type":"ContainerStarted","Data":"86111e5b851c8cb6b1e16e78c0cfc857fe872e70f4ffddcc3f699b209441f613"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.707007    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m" event={"ID":"9adf37cf-0b32-429f-ad06-819052944ee0","Type":"ContainerStarted","Data":"e9401a5bdcbd7877914d5ec34a0d721585ded112bfbe7caf6bc9bd6075af3383"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.708220    4824 generic.go:334] "Generic (PLEG): container finished" podID="eae9e020-d1a0-467f-b10b-9e022f357fc3" containerID="23f6131210ede6ea7ba593143c84c75a67839933dcb9683b76482ec317d75e0d" exitCode=0
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.708258    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp" event={"ID":"eae9e020-d1a0-467f-b10b-9e022f357fc3","Type":"ContainerDied","Data":"23f6131210ede6ea7ba593143c84c75a67839933dcb9683b76482ec317d75e0d"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.712146    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf8g8" event={"ID":"a7596399-0f46-4c4a-97ae-379efd65e7af","Type":"ContainerStarted","Data":"ae4cdd0775aebbec00684de7c4e7a9a996c7eb1789c180d983d5dabb414e162a"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.713849    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4217cd82a6ca1bf581c4cc0f199b1789928c98a0907d200185247895d6a3b9df"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.713897    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7de0ca1bd5c75373cfd0eaab059b96719640cb4975bc11ea6f7e9d4854fbb652"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.714110    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.715559    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz" event={"ID":"9c2ebdfa-baca-4839-8524-e4161b7cbe89","Type":"ContainerStarted","Data":"218050e59955e7c738a66856bcd5e861eb316a3bd90362f956ef0ad22ee68ed7"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.716370    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.725805    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zcblz" podStartSLOduration=8.725785998 podStartE2EDuration="8.725785998s" podCreationTimestamp="2026-01-22 09:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.725598151 +0000 UTC m=+156.629075144" watchObservedRunningTime="2026-01-22 09:45:23.725785998 +0000 UTC m=+156.629262991"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.739075    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk" event={"ID":"602843d5-8e62-4834-a7ae-af01d9f14644","Type":"ContainerStarted","Data":"b075f5464f77e83bd960f9f5d197fd791b89733a5025b8fc589380ed370c38e2"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.743978    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.744427    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.244400983 +0000 UTC m=+157.147877976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.747886    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x" event={"ID":"0e085a8e-5760-4859-bf6b-ac4fac6aa9cf","Type":"ContainerStarted","Data":"ede291bd62b5f14f22ab17b6a6fba1ed31ce83cac5567ae13ae1f43cb1b5dcac"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.757190    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-q4jqv" event={"ID":"697490dd-5ba8-4719-b8a3-87a9b9ed08b3","Type":"ContainerStarted","Data":"cb45976088e514fe774090ddb9ab482a7c19336a9b6a5bf0c32e231ce2b08859"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.788145    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-lprcb" event={"ID":"6803cfcb-a020-4f9c-a668-a2df6abbb407","Type":"ContainerStarted","Data":"d5e99b9d712bd5891b1f8197e5cff9f8e1ffbb6e438da5f094338361472d3cfe"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.809842    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" event={"ID":"fc660b59-eb41-4c18-84a4-ebd782752570","Type":"ContainerStarted","Data":"6e0a8d776d28a55a8220214416d4cf614c7325924f7eb84819824c6ebbcb6a02"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.829767    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m" podStartSLOduration=131.829745911 podStartE2EDuration="2m11.829745911s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.829567064 +0000 UTC m=+156.733044047" watchObservedRunningTime="2026-01-22 09:45:23.829745911 +0000 UTC m=+156.733222904"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.829790    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x" event={"ID":"a1215925-7f63-449c-8aa1-879ac01ed726","Type":"ContainerStarted","Data":"c470d9cbe2781f1945f1eb34e87de8d695ceb052e0f21782516a3fd776c6f7dc"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.844793    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rmsj9" event={"ID":"4809673b-3d53-4bd5-8739-9be6a2443e60","Type":"ContainerStarted","Data":"5089bf912cf26f4dae1f40ae8f29e297aa214a1d11ab1e7ac515f44920792ce8"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.845597    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.845991    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.848905    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.348891116 +0000 UTC m=+157.252368109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.849493    4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-rmsj9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body=
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.849532    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rmsj9" podUID="4809673b-3d53-4bd5-8739-9be6a2443e60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.872106    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"990c5aea44e0e9186750e69cd20bb325fc6c7b82095d003c2d74bcd9a05f5eda"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.876557    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:23 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:23 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:23 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.876615    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.878147    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f79tb" event={"ID":"7525c5ea-34c5-46e7-b49f-0c4a2a06e0fa","Type":"ContainerStarted","Data":"f7d38fea97ebe0d2537eaf833c936133918bf3ae6e95cc3403cf70abe5516ac7"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.889123    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz" podStartSLOduration=131.889099348 podStartE2EDuration="2m11.889099348s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.885756584 +0000 UTC m=+156.789233577" watchObservedRunningTime="2026-01-22 09:45:23.889099348 +0000 UTC m=+156.792576331"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.894536    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x7psz" event={"ID":"8d492ef8-5e48-4fc7-a56b-1233da10e9f7","Type":"ContainerStarted","Data":"0f4479d26a48390242b7c1ff7153b306b3b7aeefddac12b46b15bd8294207e9c"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.908098    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0a8e16823e2e2fe9509e32f49fdee63a28ecc7a1deb288eba40bf96a98001fea"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.908157    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"496a4c1786fc1c7877cbdefccb340cfe7004b32b62ac42ce10ff484d3c7cf7ea"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.914152    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj" event={"ID":"cb724a14-a04d-458a-8350-65e8f9d5a69e","Type":"ContainerStarted","Data":"de1ca1e05f7b6e3f101e4025eefbe9ab4a9b5a4d506edaa3ab268ecde1df1c96"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.924757    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" event={"ID":"2463610b-364e-4154-83ee-b6c00a6c6da5","Type":"ContainerStarted","Data":"98fd915035c3a2f6e9220db622a5677a3e317ca2d01d7ecb30a5377bfb45b1f2"}
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.926087    4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7qjxk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body=
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.926172    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.933529    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-29d5x" podStartSLOduration=131.933504157 podStartE2EDuration="2m11.933504157s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.928827893 +0000 UTC m=+156.832304886" watchObservedRunningTime="2026-01-22 09:45:23.933504157 +0000 UTC m=+156.836981150"
Jan 22 09:45:23 crc kubenswrapper[4824]: I0122 09:45:23.947502    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:23 crc kubenswrapper[4824]: E0122 09:45:23.948823    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.448802919 +0000 UTC m=+157.352279912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.003657    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-f79tb" podStartSLOduration=132.003626717 podStartE2EDuration="2m12.003626717s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:23.999990581 +0000 UTC m=+156.903467574" watchObservedRunningTime="2026-01-22 09:45:24.003626717 +0000 UTC m=+156.907103710"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.023161    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rgv6r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.056173    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.068116    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.568086735 +0000 UTC m=+157.471563728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.108900    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-hpp2x" podStartSLOduration=132.108875619 podStartE2EDuration="2m12.108875619s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:24.055178893 +0000 UTC m=+156.958655886" watchObservedRunningTime="2026-01-22 09:45:24.108875619 +0000 UTC m=+157.012352612"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.162786    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.163174    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.663152596 +0000 UTC m=+157.566629589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.177599    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-22 09:40:23 +0000 UTC, rotation deadline is 2026-12-05 08:22:22.132963303 +0000 UTC
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.177667    4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7606h36m57.955301015s for next certificate rotation
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.206991    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-p4xwj" podStartSLOduration=132.206960853 podStartE2EDuration="2m12.206960853s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:24.164764337 +0000 UTC m=+157.068241330" watchObservedRunningTime="2026-01-22 09:45:24.206960853 +0000 UTC m=+157.110437846"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.207311    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rmsj9" podStartSLOduration=132.207303186 podStartE2EDuration="2m12.207303186s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:24.110112145 +0000 UTC m=+157.013589138" watchObservedRunningTime="2026-01-22 09:45:24.207303186 +0000 UTC m=+157.110780189"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.265564    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.266085    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.766067021 +0000 UTC m=+157.669544014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.363917    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qvd2r"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.365507    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.374339    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.374677    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.874653577 +0000 UTC m=+157.778130570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.378976    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.379138    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.397772    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qvd2r"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.476537    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eae9e020-d1a0-467f-b10b-9e022f357fc3-secret-volume\") pod \"eae9e020-d1a0-467f-b10b-9e022f357fc3\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.476622    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eae9e020-d1a0-467f-b10b-9e022f357fc3-config-volume\") pod \"eae9e020-d1a0-467f-b10b-9e022f357fc3\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.476766    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ssf7\" (UniqueName: \"kubernetes.io/projected/eae9e020-d1a0-467f-b10b-9e022f357fc3-kube-api-access-4ssf7\") pod \"eae9e020-d1a0-467f-b10b-9e022f357fc3\" (UID: \"eae9e020-d1a0-467f-b10b-9e022f357fc3\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.476905    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-catalog-content\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.476974    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kcw2\" (UniqueName: \"kubernetes.io/projected/88bd6885-a443-4d1f-b06e-5dd0f3107dab-kube-api-access-8kcw2\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.477006    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.477045    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-utilities\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.477585    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eae9e020-d1a0-467f-b10b-9e022f357fc3-config-volume" (OuterVolumeSpecName: "config-volume") pod "eae9e020-d1a0-467f-b10b-9e022f357fc3" (UID: "eae9e020-d1a0-467f-b10b-9e022f357fc3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.477847    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:24.977823212 +0000 UTC m=+157.881300305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.516622    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae9e020-d1a0-467f-b10b-9e022f357fc3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eae9e020-d1a0-467f-b10b-9e022f357fc3" (UID: "eae9e020-d1a0-467f-b10b-9e022f357fc3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.524879    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae9e020-d1a0-467f-b10b-9e022f357fc3-kube-api-access-4ssf7" (OuterVolumeSpecName: "kube-api-access-4ssf7") pod "eae9e020-d1a0-467f-b10b-9e022f357fc3" (UID: "eae9e020-d1a0-467f-b10b-9e022f357fc3"). InnerVolumeSpecName "kube-api-access-4ssf7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.549053    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9rdhx"]
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.549306    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae9e020-d1a0-467f-b10b-9e022f357fc3" containerName="collect-profiles"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.549323    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae9e020-d1a0-467f-b10b-9e022f357fc3" containerName="collect-profiles"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.549457    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae9e020-d1a0-467f-b10b-9e022f357fc3" containerName="collect-profiles"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.550314    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.567512    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.571734    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9rdhx"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579284    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579598    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kcw2\" (UniqueName: \"kubernetes.io/projected/88bd6885-a443-4d1f-b06e-5dd0f3107dab-kube-api-access-8kcw2\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579676    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-utilities\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579720    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-catalog-content\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579768    4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eae9e020-d1a0-467f-b10b-9e022f357fc3-config-volume\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579787    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ssf7\" (UniqueName: \"kubernetes.io/projected/eae9e020-d1a0-467f-b10b-9e022f357fc3-kube-api-access-4ssf7\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.579803    4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eae9e020-d1a0-467f-b10b-9e022f357fc3-secret-volume\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.580242    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-catalog-content\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.580340    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.080316861 +0000 UTC m=+157.983793854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.580898    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-utilities\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.636030    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kcw2\" (UniqueName: \"kubernetes.io/projected/88bd6885-a443-4d1f-b06e-5dd0f3107dab-kube-api-access-8kcw2\") pod \"certified-operators-qvd2r\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") " pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.682113    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-utilities\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.682196    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-catalog-content\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.682270    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.682333    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdkr9\" (UniqueName: \"kubernetes.io/projected/072ef9f4-8754-45c5-9600-accb760d786c-kube-api-access-xdkr9\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.682851    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.18283558 +0000 UTC m=+158.086312573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.704816    4824 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wgvg2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body=
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.704897    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2" podUID="97038c7c-c8e0-451d-86a5-93e5f36d5e33" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.716138    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.746205    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nj5jb"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.747524    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.779619    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nj5jb"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.787702    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.787861    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-catalog-content\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.787933    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdkr9\" (UniqueName: \"kubernetes.io/projected/072ef9f4-8754-45c5-9600-accb760d786c-kube-api-access-xdkr9\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.787969    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-utilities\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.788348    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-utilities\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.788483    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-catalog-content\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.788612    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.28856597 +0000 UTC m=+158.192042963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.819376    4824 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-822xl container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.27:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body=
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.819470    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.27:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.830130    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdkr9\" (UniqueName: \"kubernetes.io/projected/072ef9f4-8754-45c5-9600-accb760d786c-kube-api-access-xdkr9\") pod \"community-operators-9rdhx\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") " pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.831820    4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.880558    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:24 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:24 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:24 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.880640    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.889398    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-catalog-content\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.889673    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-utilities\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.889698    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjbk6\" (UniqueName: \"kubernetes.io/projected/1ddf148c-bafe-480f-b156-e3daa798bb31-kube-api-access-rjbk6\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.889750    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.890177    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.390161914 +0000 UTC m=+158.293638907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.948704    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.948690    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484585-7rhkp" event={"ID":"eae9e020-d1a0-467f-b10b-9e022f357fc3","Type":"ContainerDied","Data":"5c37d5229908deafd9dfb7877a5bd261fd62b4b8d2e6044d8974d7ed605e2c1d"}
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.948831    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c37d5229908deafd9dfb7877a5bd261fd62b4b8d2e6044d8974d7ed605e2c1d"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.951871    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd" event={"ID":"c8b117b7-e048-4a4a-9592-1d3eb821d309","Type":"ContainerStarted","Data":"950c46e14fec7904ed98d5f76250610a966dd5c620d310b441ec3b3da87b3770"}
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.958202    4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-rmsj9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body=
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.958273    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rmsj9" podUID="4809673b-3d53-4bd5-8739-9be6a2443e60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.964799    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.968725    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wlq2"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.969926    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.978677    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.985488    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wlq2"]
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.993039    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.993267    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-utilities\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.993297    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjbk6\" (UniqueName: \"kubernetes.io/projected/1ddf148c-bafe-480f-b156-e3daa798bb31-kube-api-access-rjbk6\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.993411    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-catalog-content\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.993879    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-catalog-content\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:24 crc kubenswrapper[4824]: E0122 09:45:24.994987    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.494937969 +0000 UTC m=+158.398414962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:24 crc kubenswrapper[4824]: I0122 09:45:24.995214    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-utilities\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.021343    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjbk6\" (UniqueName: \"kubernetes.io/projected/1ddf148c-bafe-480f-b156-e3daa798bb31-kube-api-access-rjbk6\") pod \"certified-operators-nj5jb\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") " pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.096432    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trrnh\" (UniqueName: \"kubernetes.io/projected/2fd47eee-b6b0-4448-8bcd-09c188762714-kube-api-access-trrnh\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.096781    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-catalog-content\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.097040    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-utilities\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.097062    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.107444    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.607335028 +0000 UTC m=+158.510812021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.112785    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wgvg2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.123316    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.194784    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.197868    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.198093    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trrnh\" (UniqueName: \"kubernetes.io/projected/2fd47eee-b6b0-4448-8bcd-09c188762714-kube-api-access-trrnh\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.198139    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-catalog-content\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.198176    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-utilities\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.199144    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-utilities\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.199249    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.69922665 +0000 UTC m=+158.602703643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.199760    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-catalog-content\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.246634    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trrnh\" (UniqueName: \"kubernetes.io/projected/2fd47eee-b6b0-4448-8bcd-09c188762714-kube-api-access-trrnh\") pod \"community-operators-9wlq2\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") " pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.295662    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.300845    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.301317    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.801302704 +0000 UTC m=+158.704779697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.401845    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.402016    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.901986895 +0000 UTC m=+158.805463888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.402128    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.402554    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:25.902543986 +0000 UTC m=+158.806020979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.445666    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qvd2r"]
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.503240    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.503636    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-22 09:45:26.003613611 +0000 UTC m=+158.907090604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.571447    4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-22T09:45:24.831856577Z","Handler":null,"Name":""}
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.604941    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: E0122 09:45:25.605418    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-22 09:45:26.105405014 +0000 UTC m=+159.008882007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zbdgw" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.617633    4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.617684    4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.706573    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") "
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.713277    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9rdhx"]
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.750052    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue ""
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.810007    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.822668    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nj5jb"]
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.826141    4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice...
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.826243    4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.877216    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:25 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:25 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:25 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.877284    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.936858    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zbdgw\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") " pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.975484    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd" event={"ID":"c8b117b7-e048-4a4a-9592-1d3eb821d309","Type":"ContainerStarted","Data":"d4b464e9fe147742bca5ab7f83479bde72cd105d254ba2bbd3577196ac1e4e37"}
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.988826    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nj5jb" event={"ID":"1ddf148c-bafe-480f-b156-e3daa798bb31","Type":"ContainerStarted","Data":"46f5ab80d4eb8e567583e991512a1764cf52efd6f6abac59bae64941536406ce"}
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.990457    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wlq2"]
Jan 22 09:45:25 crc kubenswrapper[4824]: I0122 09:45:25.993439    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rdhx" event={"ID":"072ef9f4-8754-45c5-9600-accb760d786c","Type":"ContainerStarted","Data":"a66c02577ce5c9068736f7d9f877b8f61e9940da717b3693ce322dd07db5d237"}
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.005782    4824 generic.go:334] "Generic (PLEG): container finished" podID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerID="11133aa1bdf4f62b5e6e2961b375b90475d985dd21a6ecbc90f2326ffca0d637" exitCode=0
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.005988    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qvd2r" event={"ID":"88bd6885-a443-4d1f-b06e-5dd0f3107dab","Type":"ContainerDied","Data":"11133aa1bdf4f62b5e6e2961b375b90475d985dd21a6ecbc90f2326ffca0d637"}
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.006023    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qvd2r" event={"ID":"88bd6885-a443-4d1f-b06e-5dd0f3107dab","Type":"ContainerStarted","Data":"0bc12fab86f9e0a51fcf153c6a77997b661babe9666acb1fd1935a94d134989c"}
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.006869    4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-rmsj9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body=
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.006907    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rmsj9" podUID="4809673b-3d53-4bd5-8739-9be6a2443e60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.007202    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-hmgmd" podStartSLOduration=11.007181753 podStartE2EDuration="11.007181753s" podCreationTimestamp="2026-01-22 09:45:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:26.000458642 +0000 UTC m=+158.903935635" watchObservedRunningTime="2026-01-22 09:45:26.007181753 +0000 UTC m=+158.910658746"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.009479    4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.018230    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.316778    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbdgw"]
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.331635    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xfqfs"]
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.332729    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.335387    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.346739    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfqfs"]
Jan 22 09:45:26 crc kubenswrapper[4824]: W0122 09:45:26.351325    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9633603d_c8b9_4f94_a3c5_3d882308a586.slice/crio-73438183f6f933033b7002f1b6135147cae0411ee1dd6fe5d90fc317f68177d1 WatchSource:0}: Error finding container 73438183f6f933033b7002f1b6135147cae0411ee1dd6fe5d90fc317f68177d1: Status 404 returned error can't find the container with id 73438183f6f933033b7002f1b6135147cae0411ee1dd6fe5d90fc317f68177d1
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.421636    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nbgs\" (UniqueName: \"kubernetes.io/projected/8d2678ef-e863-4181-a497-b98b7fff0cce-kube-api-access-9nbgs\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.421693    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-catalog-content\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.421743    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-utilities\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.477746    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-b5fvk"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.523152    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nbgs\" (UniqueName: \"kubernetes.io/projected/8d2678ef-e863-4181-a497-b98b7fff0cce-kube-api-access-9nbgs\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.523206    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-catalog-content\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.523244    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-utilities\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.523704    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-utilities\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.523780    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-catalog-content\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.547745    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nbgs\" (UniqueName: \"kubernetes.io/projected/8d2678ef-e863-4181-a497-b98b7fff0cce-kube-api-access-9nbgs\") pod \"redhat-marketplace-xfqfs\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") " pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.661312    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.734447    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hv6zw"]
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.735461    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.748594    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hv6zw"]
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.826205    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-catalog-content\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.826252    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-utilities\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.826284    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8phnl\" (UniqueName: \"kubernetes.io/projected/cfe569b9-ef9e-4ce9-b985-511797cc0633-kube-api-access-8phnl\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.872592    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:26 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:26 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:26 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.872660    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.927633    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8phnl\" (UniqueName: \"kubernetes.io/projected/cfe569b9-ef9e-4ce9-b985-511797cc0633-kube-api-access-8phnl\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.927756    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-catalog-content\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.927790    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-utilities\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.928301    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-catalog-content\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.928382    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-utilities\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:26 crc kubenswrapper[4824]: I0122 09:45:26.944654    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8phnl\" (UniqueName: \"kubernetes.io/projected/cfe569b9-ef9e-4ce9-b985-511797cc0633-kube-api-access-8phnl\") pod \"redhat-marketplace-hv6zw\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") " pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.012570    4824 generic.go:334] "Generic (PLEG): container finished" podID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerID="891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e" exitCode=0
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.012658    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nj5jb" event={"ID":"1ddf148c-bafe-480f-b156-e3daa798bb31","Type":"ContainerDied","Data":"891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e"}
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.015113    4824 generic.go:334] "Generic (PLEG): container finished" podID="072ef9f4-8754-45c5-9600-accb760d786c" containerID="ef1771a84cf61915d4b0d975f9ee8872b5cd82196ad2083f1dd983ba605b8a06" exitCode=0
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.015192    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rdhx" event={"ID":"072ef9f4-8754-45c5-9600-accb760d786c","Type":"ContainerDied","Data":"ef1771a84cf61915d4b0d975f9ee8872b5cd82196ad2083f1dd983ba605b8a06"}
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.019064    4824 generic.go:334] "Generic (PLEG): container finished" podID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerID="d2bddba741f540f51d957b8348722b2dd5b6390ad7a31df14141f87d367c925f" exitCode=0
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.019157    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerDied","Data":"d2bddba741f540f51d957b8348722b2dd5b6390ad7a31df14141f87d367c925f"}
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.019223    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerStarted","Data":"e106e71f1ac6b3af543619475b1236acef177c6ac8f4ee29cb9d38d8fb81440c"}
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.021054    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" event={"ID":"9633603d-c8b9-4f94-a3c5-3d882308a586","Type":"ContainerStarted","Data":"8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd"}
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.021081    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" event={"ID":"9633603d-c8b9-4f94-a3c5-3d882308a586","Type":"ContainerStarted","Data":"73438183f6f933033b7002f1b6135147cae0411ee1dd6fe5d90fc317f68177d1"}
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.021664    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.059738    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" podStartSLOduration=135.059715153 podStartE2EDuration="2m15.059715153s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:27.056591307 +0000 UTC m=+159.960068330" watchObservedRunningTime="2026-01-22 09:45:27.059715153 +0000 UTC m=+159.963192146"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.068938    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.126018    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfqfs"]
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.311724    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hv6zw"]
Jan 22 09:45:27 crc kubenswrapper[4824]: W0122 09:45:27.322679    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfe569b9_ef9e_4ce9_b985_511797cc0633.slice/crio-af979d858887a89032987205e217b373121c24078b56810ce5bc4a6ff2fda3ec WatchSource:0}: Error finding container af979d858887a89032987205e217b373121c24078b56810ce5bc4a6ff2fda3ec: Status 404 returned error can't find the container with id af979d858887a89032987205e217b373121c24078b56810ce5bc4a6ff2fda3ec
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.412685    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.433063    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.433137    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.434513    4824 patch_prober.go:28] interesting pod/console-f9d7485db-8878r container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body=
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.434567    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8878r" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.581953    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.582278    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.587057    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.757555    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bkmd8"]
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.759197    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.761830    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.777264    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bkmd8"]
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.838432    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88zfd\" (UniqueName: \"kubernetes.io/projected/2b498256-7f0d-4941-998e-e670f5941cdb-kube-api-access-88zfd\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.838584    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-utilities\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.838736    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-catalog-content\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.869210    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.871996    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:27 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:27 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:27 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.872041    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.940066    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-catalog-content\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.940180    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88zfd\" (UniqueName: \"kubernetes.io/projected/2b498256-7f0d-4941-998e-e670f5941cdb-kube-api-access-88zfd\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.940293    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-utilities\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.940662    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-catalog-content\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.940721    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-utilities\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:27 crc kubenswrapper[4824]: I0122 09:45:27.959672    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88zfd\" (UniqueName: \"kubernetes.io/projected/2b498256-7f0d-4941-998e-e670f5941cdb-kube-api-access-88zfd\") pod \"redhat-operators-bkmd8\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") " pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.026517    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfqfs" event={"ID":"8d2678ef-e863-4181-a497-b98b7fff0cce","Type":"ContainerStarted","Data":"68395353640e3ee79eca6e1f24ad37e6e0a60966ea8e738e92a9af65c677747b"}
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.027543    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hv6zw" event={"ID":"cfe569b9-ef9e-4ce9-b985-511797cc0633","Type":"ContainerStarted","Data":"af979d858887a89032987205e217b373121c24078b56810ce5bc4a6ff2fda3ec"}
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.032593    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-mbh2m"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.086370    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.171095    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.172836    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.180224    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"]
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.181288    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.186377    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"]
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.201682    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.201676    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.229316    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.349755    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgxdf\" (UniqueName: \"kubernetes.io/projected/05d38c4c-ca18-478a-bcd5-01800ded772a-kube-api-access-qgxdf\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.350083    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-catalog-content\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.350115    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.350149    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-utilities\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.350197    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: E0122 09:45:28.401046    4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfe569b9_ef9e_4ce9_b985_511797cc0633.slice/crio-conmon-cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05.scope\": RecentStats: unable to find data in memory cache]"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.451250    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-catalog-content\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.451316    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.451376    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-utilities\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.451420    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.451442    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgxdf\" (UniqueName: \"kubernetes.io/projected/05d38c4c-ca18-478a-bcd5-01800ded772a-kube-api-access-qgxdf\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.451590    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.452230    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-catalog-content\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.452282    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-utilities\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.469124    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgxdf\" (UniqueName: \"kubernetes.io/projected/05d38c4c-ca18-478a-bcd5-01800ded772a-kube-api-access-qgxdf\") pod \"redhat-operators-xfnlv\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") " pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.471754    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.522560    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.543638    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.560538    4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-rmsj9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body=
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.560610    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rmsj9" podUID="4809673b-3d53-4bd5-8739-9be6a2443e60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.560925    4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-rmsj9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body=
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.561013    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rmsj9" podUID="4809673b-3d53-4bd5-8739-9be6a2443e60" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.586114    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bkmd8"]
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.874030    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:28 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:28 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:28 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.874464    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:28 crc kubenswrapper[4824]: I0122 09:45:28.893555    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:45:28 crc kubenswrapper[4824]: W0122 09:45:28.914193    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05d38c4c_ca18_478a_bcd5_01800ded772a.slice/crio-bdc553235dfbef374e8a036f93097d785519ebdd499edd1a57480c3b17ce406a WatchSource:0}: Error finding container bdc553235dfbef374e8a036f93097d785519ebdd499edd1a57480c3b17ce406a: Status 404 returned error can't find the container with id bdc553235dfbef374e8a036f93097d785519ebdd499edd1a57480c3b17ce406a
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.033660    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"]
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.040215    4824 generic.go:334] "Generic (PLEG): container finished" podID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerID="cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05" exitCode=0
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.040284    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hv6zw" event={"ID":"cfe569b9-ef9e-4ce9-b985-511797cc0633","Type":"ContainerDied","Data":"cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05"}
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.042267    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerStarted","Data":"bdc553235dfbef374e8a036f93097d785519ebdd499edd1a57480c3b17ce406a"}
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.047080    4824 generic.go:334] "Generic (PLEG): container finished" podID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerID="5f4a46dd06222cb576b91f5e069fc6c73acbbfefc9b525b8a75e4ea71d2dc2ba" exitCode=0
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.047118    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfqfs" event={"ID":"8d2678ef-e863-4181-a497-b98b7fff0cce","Type":"ContainerDied","Data":"5f4a46dd06222cb576b91f5e069fc6c73acbbfefc9b525b8a75e4ea71d2dc2ba"}
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.054255    4824 generic.go:334] "Generic (PLEG): container finished" podID="2b498256-7f0d-4941-998e-e670f5941cdb" containerID="02289cdb515a3089d234977eda3d52348ec5a570a77a9632cf626c40ff2ca828" exitCode=0
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.054406    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerDied","Data":"02289cdb515a3089d234977eda3d52348ec5a570a77a9632cf626c40ff2ca828"}
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.057697    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerStarted","Data":"89aebd8e5cf102fe49d7ad6910519189445e61b94c44aac76cb2dd657f817639"}
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.872234    4824 patch_prober.go:28] interesting pod/router-default-5444994796-b8244 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld
Jan 22 09:45:29 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld
Jan 22 09:45:29 crc kubenswrapper[4824]: [+]process-running ok
Jan 22 09:45:29 crc kubenswrapper[4824]: healthz check failed
Jan 22 09:45:29 crc kubenswrapper[4824]: I0122 09:45:29.872530    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-b8244" podUID="12dcf185-540b-4a55-89e6-cbc3b904a15c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500"
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.089467    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e557a328-3f83-40c0-919b-4fd30e2fc8ab","Type":"ContainerStarted","Data":"e3b4db5f582b0fb6ed99c174aae8c542d9a403a304afd76a065c6e82e2719929"}
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.089572    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e557a328-3f83-40c0-919b-4fd30e2fc8ab","Type":"ContainerStarted","Data":"675fa4a05ce4cb6a4de0ba911ba8798b04331817a94279d6d084ab4da0015100"}
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.092202    4824 generic.go:334] "Generic (PLEG): container finished" podID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerID="818ce7d2540b4e6464a5ff26be6cb503b5edb5c2cddfd5c3d68aab18c0efe0fa" exitCode=0
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.092251    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerDied","Data":"818ce7d2540b4e6464a5ff26be6cb503b5edb5c2cddfd5c3d68aab18c0efe0fa"}
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.104739    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.104722617 podStartE2EDuration="2.104722617s" podCreationTimestamp="2026-01-22 09:45:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:30.103931487 +0000 UTC m=+163.007408490" watchObservedRunningTime="2026-01-22 09:45:30.104722617 +0000 UTC m=+163.008199610"
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.332690    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.332815    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.871729    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:30 crc kubenswrapper[4824]: I0122 09:45:30.876417    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-b8244"
Jan 22 09:45:31 crc kubenswrapper[4824]: I0122 09:45:31.207162    4824 generic.go:334] "Generic (PLEG): container finished" podID="e557a328-3f83-40c0-919b-4fd30e2fc8ab" containerID="e3b4db5f582b0fb6ed99c174aae8c542d9a403a304afd76a065c6e82e2719929" exitCode=0
Jan 22 09:45:31 crc kubenswrapper[4824]: I0122 09:45:31.207291    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e557a328-3f83-40c0-919b-4fd30e2fc8ab","Type":"ContainerDied","Data":"e3b4db5f582b0fb6ed99c174aae8c542d9a403a304afd76a065c6e82e2719929"}
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.657040    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.749496    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kube-api-access\") pod \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") "
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.749558    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kubelet-dir\") pod \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\" (UID: \"e557a328-3f83-40c0-919b-4fd30e2fc8ab\") "
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.749650    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e557a328-3f83-40c0-919b-4fd30e2fc8ab" (UID: "e557a328-3f83-40c0-919b-4fd30e2fc8ab"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.750060    4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kubelet-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.777572    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e557a328-3f83-40c0-919b-4fd30e2fc8ab" (UID: "e557a328-3f83-40c0-919b-4fd30e2fc8ab"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:45:32 crc kubenswrapper[4824]: I0122 09:45:32.851050    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e557a328-3f83-40c0-919b-4fd30e2fc8ab-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.237224    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e557a328-3f83-40c0-919b-4fd30e2fc8ab","Type":"ContainerDied","Data":"675fa4a05ce4cb6a4de0ba911ba8798b04331817a94279d6d084ab4da0015100"}
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.237266    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="675fa4a05ce4cb6a4de0ba911ba8798b04331817a94279d6d084ab4da0015100"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.237282    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.376208    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"]
Jan 22 09:45:33 crc kubenswrapper[4824]: E0122 09:45:33.376540    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e557a328-3f83-40c0-919b-4fd30e2fc8ab" containerName="pruner"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.376556    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e557a328-3f83-40c0-919b-4fd30e2fc8ab" containerName="pruner"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.376728    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e557a328-3f83-40c0-919b-4fd30e2fc8ab" containerName="pruner"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.377498    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.384448    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.384700    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.394107    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"]
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.397609    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zcblz"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.559409    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77f348cc-22dc-41c0-a184-98721575c459-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") " pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.564833    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77f348cc-22dc-41c0-a184-98721575c459-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") " pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.667341    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77f348cc-22dc-41c0-a184-98721575c459-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") " pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.667899    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77f348cc-22dc-41c0-a184-98721575c459-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") " pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.668342    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77f348cc-22dc-41c0-a184-98721575c459-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") " pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.690802    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77f348cc-22dc-41c0-a184-98721575c459-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") " pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:33 crc kubenswrapper[4824]: I0122 09:45:33.718595    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:34 crc kubenswrapper[4824]: I0122 09:45:34.292836    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"]
Jan 22 09:45:34 crc kubenswrapper[4824]: I0122 09:45:34.896517    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:34 crc kubenswrapper[4824]: I0122 09:45:34.904098    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f43b4618-291a-40d3-b1ca-32abfadaf376-metrics-certs\") pod \"network-metrics-daemon-xxtjz\" (UID: \"f43b4618-291a-40d3-b1ca-32abfadaf376\") " pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:35 crc kubenswrapper[4824]: I0122 09:45:35.133233    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xxtjz"
Jan 22 09:45:35 crc kubenswrapper[4824]: I0122 09:45:35.264175    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"77f348cc-22dc-41c0-a184-98721575c459","Type":"ContainerStarted","Data":"04c714073f32e7bb6b215837929b9e27c273fd1baaaf4afbc03c1c9766c29d97"}
Jan 22 09:45:35 crc kubenswrapper[4824]: I0122 09:45:35.698260    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xxtjz"]
Jan 22 09:45:35 crc kubenswrapper[4824]: W0122 09:45:35.720663    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf43b4618_291a_40d3_b1ca_32abfadaf376.slice/crio-7b30d9df9ac497826ad141dbdaf87d3c830c6d2d0992a07a21dd8da2bbe499a8 WatchSource:0}: Error finding container 7b30d9df9ac497826ad141dbdaf87d3c830c6d2d0992a07a21dd8da2bbe499a8: Status 404 returned error can't find the container with id 7b30d9df9ac497826ad141dbdaf87d3c830c6d2d0992a07a21dd8da2bbe499a8
Jan 22 09:45:36 crc kubenswrapper[4824]: I0122 09:45:36.271013    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"77f348cc-22dc-41c0-a184-98721575c459","Type":"ContainerStarted","Data":"b16376ca35cf255ecc66f019af9b960668770712d334ab35d6cd15ef2f2ea7c3"}
Jan 22 09:45:36 crc kubenswrapper[4824]: I0122 09:45:36.272348    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" event={"ID":"f43b4618-291a-40d3-b1ca-32abfadaf376","Type":"ContainerStarted","Data":"7b30d9df9ac497826ad141dbdaf87d3c830c6d2d0992a07a21dd8da2bbe499a8"}
Jan 22 09:45:36 crc kubenswrapper[4824]: I0122 09:45:36.292801    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.292775575 podStartE2EDuration="3.292775575s" podCreationTimestamp="2026-01-22 09:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:45:36.283030391 +0000 UTC m=+169.186507394" watchObservedRunningTime="2026-01-22 09:45:36.292775575 +0000 UTC m=+169.196252568"
Jan 22 09:45:37 crc kubenswrapper[4824]: I0122 09:45:37.283439    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" event={"ID":"f43b4618-291a-40d3-b1ca-32abfadaf376","Type":"ContainerStarted","Data":"1199a1b98ab86f91f3fdda5242a806dcad09932bc7bdeb7ca64b0d496502c76f"}
Jan 22 09:45:37 crc kubenswrapper[4824]: I0122 09:45:37.286167    4824 generic.go:334] "Generic (PLEG): container finished" podID="77f348cc-22dc-41c0-a184-98721575c459" containerID="b16376ca35cf255ecc66f019af9b960668770712d334ab35d6cd15ef2f2ea7c3" exitCode=0
Jan 22 09:45:37 crc kubenswrapper[4824]: I0122 09:45:37.286195    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"77f348cc-22dc-41c0-a184-98721575c459","Type":"ContainerDied","Data":"b16376ca35cf255ecc66f019af9b960668770712d334ab35d6cd15ef2f2ea7c3"}
Jan 22 09:45:37 crc kubenswrapper[4824]: I0122 09:45:37.433837    4824 patch_prober.go:28] interesting pod/console-f9d7485db-8878r container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body=
Jan 22 09:45:37 crc kubenswrapper[4824]: I0122 09:45:37.433898    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8878r" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused"
Jan 22 09:45:38 crc kubenswrapper[4824]: I0122 09:45:38.570431    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rmsj9"
Jan 22 09:45:40 crc kubenswrapper[4824]: I0122 09:45:40.161316    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xk8qh"]
Jan 22 09:45:40 crc kubenswrapper[4824]: I0122 09:45:40.161831    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh" podUID="88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" containerName="controller-manager" containerID="cri-o://8635b27ca4ff819e02855846c1c473131753df1e9c83d5a4b11d210e39b2e430" gracePeriod=30
Jan 22 09:45:40 crc kubenswrapper[4824]: I0122 09:45:40.179802    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"]
Jan 22 09:45:40 crc kubenswrapper[4824]: I0122 09:45:40.180055    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" podUID="78d695a9-430a-4540-8a70-5c2bcfc40b34" containerName="route-controller-manager" containerID="cri-o://25bbcfb7b006dc80251509eba288add7fa18779d72388522ce04ae6d15473b35" gracePeriod=30
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.336909    4824 generic.go:334] "Generic (PLEG): container finished" podID="78d695a9-430a-4540-8a70-5c2bcfc40b34" containerID="25bbcfb7b006dc80251509eba288add7fa18779d72388522ce04ae6d15473b35" exitCode=0
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.337034    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" event={"ID":"78d695a9-430a-4540-8a70-5c2bcfc40b34","Type":"ContainerDied","Data":"25bbcfb7b006dc80251509eba288add7fa18779d72388522ce04ae6d15473b35"}
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.340087    4824 generic.go:334] "Generic (PLEG): container finished" podID="88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" containerID="8635b27ca4ff819e02855846c1c473131753df1e9c83d5a4b11d210e39b2e430" exitCode=0
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.340139    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh" event={"ID":"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f","Type":"ContainerDied","Data":"8635b27ca4ff819e02855846c1c473131753df1e9c83d5a4b11d210e39b2e430"}
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.837663    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.942403    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77f348cc-22dc-41c0-a184-98721575c459-kube-api-access\") pod \"77f348cc-22dc-41c0-a184-98721575c459\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") "
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.942530    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77f348cc-22dc-41c0-a184-98721575c459-kubelet-dir\") pod \"77f348cc-22dc-41c0-a184-98721575c459\" (UID: \"77f348cc-22dc-41c0-a184-98721575c459\") "
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.942597    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77f348cc-22dc-41c0-a184-98721575c459-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "77f348cc-22dc-41c0-a184-98721575c459" (UID: "77f348cc-22dc-41c0-a184-98721575c459"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.942995    4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/77f348cc-22dc-41c0-a184-98721575c459-kubelet-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:43 crc kubenswrapper[4824]: I0122 09:45:43.950429    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f348cc-22dc-41c0-a184-98721575c459-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "77f348cc-22dc-41c0-a184-98721575c459" (UID: "77f348cc-22dc-41c0-a184-98721575c459"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:45:44 crc kubenswrapper[4824]: I0122 09:45:44.043897    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/77f348cc-22dc-41c0-a184-98721575c459-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:44 crc kubenswrapper[4824]: I0122 09:45:44.349057    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"77f348cc-22dc-41c0-a184-98721575c459","Type":"ContainerDied","Data":"04c714073f32e7bb6b215837929b9e27c273fd1baaaf4afbc03c1c9766c29d97"}
Jan 22 09:45:44 crc kubenswrapper[4824]: I0122 09:45:44.349676    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04c714073f32e7bb6b215837929b9e27c273fd1baaaf4afbc03c1c9766c29d97"
Jan 22 09:45:44 crc kubenswrapper[4824]: I0122 09:45:44.349170    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.603890    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.616090    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665420    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-serving-cert\") pod \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665485    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-client-ca\") pod \"78d695a9-430a-4540-8a70-5c2bcfc40b34\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665508    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-client-ca\") pod \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665573    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-config\") pod \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665612    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4dns\" (UniqueName: \"kubernetes.io/projected/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-kube-api-access-d4dns\") pod \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665640    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9n59\" (UniqueName: \"kubernetes.io/projected/78d695a9-430a-4540-8a70-5c2bcfc40b34-kube-api-access-p9n59\") pod \"78d695a9-430a-4540-8a70-5c2bcfc40b34\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665666    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-config\") pod \"78d695a9-430a-4540-8a70-5c2bcfc40b34\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665870    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78d695a9-430a-4540-8a70-5c2bcfc40b34-serving-cert\") pod \"78d695a9-430a-4540-8a70-5c2bcfc40b34\" (UID: \"78d695a9-430a-4540-8a70-5c2bcfc40b34\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.665903    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-proxy-ca-bundles\") pod \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\" (UID: \"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f\") "
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.666266    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-client-ca" (OuterVolumeSpecName: "client-ca") pod "78d695a9-430a-4540-8a70-5c2bcfc40b34" (UID: "78d695a9-430a-4540-8a70-5c2bcfc40b34"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.666804    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" (UID: "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.666884    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-config" (OuterVolumeSpecName: "config") pod "78d695a9-430a-4540-8a70-5c2bcfc40b34" (UID: "78d695a9-430a-4540-8a70-5c2bcfc40b34"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.666986    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-client-ca" (OuterVolumeSpecName: "client-ca") pod "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" (UID: "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.666970    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.667273    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.667369    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-config" (OuterVolumeSpecName: "config") pod "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" (UID: "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.670951    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78d695a9-430a-4540-8a70-5c2bcfc40b34-kube-api-access-p9n59" (OuterVolumeSpecName: "kube-api-access-p9n59") pod "78d695a9-430a-4540-8a70-5c2bcfc40b34" (UID: "78d695a9-430a-4540-8a70-5c2bcfc40b34"). InnerVolumeSpecName "kube-api-access-p9n59". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.671022    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" (UID: "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.678617    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78d695a9-430a-4540-8a70-5c2bcfc40b34-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "78d695a9-430a-4540-8a70-5c2bcfc40b34" (UID: "78d695a9-430a-4540-8a70-5c2bcfc40b34"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.678751    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-kube-api-access-d4dns" (OuterVolumeSpecName: "kube-api-access-d4dns") pod "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" (UID: "88aa9d3e-db4d-4d92-b28d-d2f12dffb87f"). InnerVolumeSpecName "kube-api-access-d4dns". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768659    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768721    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768734    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768749    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9n59\" (UniqueName: \"kubernetes.io/projected/78d695a9-430a-4540-8a70-5c2bcfc40b34-kube-api-access-p9n59\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768766    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4dns\" (UniqueName: \"kubernetes.io/projected/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f-kube-api-access-d4dns\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768778    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78d695a9-430a-4540-8a70-5c2bcfc40b34-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.768788    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78d695a9-430a-4540-8a70-5c2bcfc40b34-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.936546    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"]
Jan 22 09:45:45 crc kubenswrapper[4824]: E0122 09:45:45.936864    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" containerName="controller-manager"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.936883    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" containerName="controller-manager"
Jan 22 09:45:45 crc kubenswrapper[4824]: E0122 09:45:45.936898    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f348cc-22dc-41c0-a184-98721575c459" containerName="pruner"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.936907    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f348cc-22dc-41c0-a184-98721575c459" containerName="pruner"
Jan 22 09:45:45 crc kubenswrapper[4824]: E0122 09:45:45.936929    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78d695a9-430a-4540-8a70-5c2bcfc40b34" containerName="route-controller-manager"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.936937    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="78d695a9-430a-4540-8a70-5c2bcfc40b34" containerName="route-controller-manager"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.937051    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" containerName="controller-manager"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.937065    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f348cc-22dc-41c0-a184-98721575c459" containerName="pruner"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.937079    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="78d695a9-430a-4540-8a70-5c2bcfc40b34" containerName="route-controller-manager"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.938138    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.939165    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"]
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.941050    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.948899    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"]
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.952818    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"]
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977609    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-config\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977650    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-client-ca\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977677    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpz4s\" (UniqueName: \"kubernetes.io/projected/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-kube-api-access-lpz4s\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977696    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-config\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977710    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec14957-e115-4061-9db2-b1dce68f76ad-serving-cert\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977737    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bvtp\" (UniqueName: \"kubernetes.io/projected/7ec14957-e115-4061-9db2-b1dce68f76ad-kube-api-access-9bvtp\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977904    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-serving-cert\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977929    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-client-ca\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:45 crc kubenswrapper[4824]: I0122 09:45:45.977966    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-proxy-ca-bundles\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.026039    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079087    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpz4s\" (UniqueName: \"kubernetes.io/projected/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-kube-api-access-lpz4s\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079144    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-config\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079187    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec14957-e115-4061-9db2-b1dce68f76ad-serving-cert\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079292    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bvtp\" (UniqueName: \"kubernetes.io/projected/7ec14957-e115-4061-9db2-b1dce68f76ad-kube-api-access-9bvtp\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079374    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-serving-cert\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079420    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-client-ca\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079452    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-proxy-ca-bundles\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079573    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-config\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.079610    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-client-ca\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.081056    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-client-ca\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.082745    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-client-ca\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.084087    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-config\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.097725    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-config\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.098039    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-proxy-ca-bundles\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.101411    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec14957-e115-4061-9db2-b1dce68f76ad-serving-cert\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.101951    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-serving-cert\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.104349    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bvtp\" (UniqueName: \"kubernetes.io/projected/7ec14957-e115-4061-9db2-b1dce68f76ad-kube-api-access-9bvtp\") pod \"controller-manager-5d9b8d9bd5-b6lls\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") " pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.106651    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpz4s\" (UniqueName: \"kubernetes.io/projected/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-kube-api-access-lpz4s\") pod \"route-controller-manager-78f64778cd-v5lt4\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") " pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.280557    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.321335    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.360469    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh" event={"ID":"88aa9d3e-db4d-4d92-b28d-d2f12dffb87f","Type":"ContainerDied","Data":"69b484aadce347cd795e85bd7fc30073ba4f8c1895a209449b28438e12c54491"}
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.360515    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xk8qh"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.360537    4824 scope.go:117] "RemoveContainer" containerID="8635b27ca4ff819e02855846c1c473131753df1e9c83d5a4b11d210e39b2e430"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.363940    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc" event={"ID":"78d695a9-430a-4540-8a70-5c2bcfc40b34","Type":"ContainerDied","Data":"f733e0775cf67dc06804f2a027c1940077b00daadd8a5a8c1c922457a63de6b0"}
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.363985    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.390850    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xk8qh"]
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.394352    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xk8qh"]
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.403821    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"]
Jan 22 09:45:46 crc kubenswrapper[4824]: I0122 09:45:46.408266    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5n9hc"]
Jan 22 09:45:47 crc kubenswrapper[4824]: I0122 09:45:47.411044    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78d695a9-430a-4540-8a70-5c2bcfc40b34" path="/var/lib/kubelet/pods/78d695a9-430a-4540-8a70-5c2bcfc40b34/volumes"
Jan 22 09:45:47 crc kubenswrapper[4824]: I0122 09:45:47.411898    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88aa9d3e-db4d-4d92-b28d-d2f12dffb87f" path="/var/lib/kubelet/pods/88aa9d3e-db4d-4d92-b28d-d2f12dffb87f/volumes"
Jan 22 09:45:47 crc kubenswrapper[4824]: I0122 09:45:47.437023    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:47 crc kubenswrapper[4824]: I0122 09:45:47.441968    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:45:58 crc kubenswrapper[4824]: I0122 09:45:58.240735    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-877dz"
Jan 22 09:46:00 crc kubenswrapper[4824]: I0122 09:46:00.106156    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"]
Jan 22 09:46:00 crc kubenswrapper[4824]: I0122 09:46:00.160310    4824 scope.go:117] "RemoveContainer" containerID="25bbcfb7b006dc80251509eba288add7fa18779d72388522ce04ae6d15473b35"
Jan 22 09:46:00 crc kubenswrapper[4824]: I0122 09:46:00.199843    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"]
Jan 22 09:46:00 crc kubenswrapper[4824]: I0122 09:46:00.333522    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:46:00 crc kubenswrapper[4824]: I0122 09:46:00.333591    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:46:00 crc kubenswrapper[4824]: I0122 09:46:00.732215    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c"
Jan 22 09:46:00 crc kubenswrapper[4824]: E0122 09:46:00.805913    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18"
Jan 22 09:46:00 crc kubenswrapper[4824]: E0122 09:46:00.806172    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8kcw2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qvd2r_openshift-marketplace(88bd6885-a443-4d1f-b06e-5dd0f3107dab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:00 crc kubenswrapper[4824]: E0122 09:46:00.807334    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qvd2r" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab"
Jan 22 09:46:03 crc kubenswrapper[4824]: E0122 09:46:03.216528    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18"
Jan 22 09:46:03 crc kubenswrapper[4824]: E0122 09:46:03.216725    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rjbk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nj5jb_openshift-marketplace(1ddf148c-bafe-480f-b156-e3daa798bb31): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:03 crc kubenswrapper[4824]: E0122 09:46:03.217967    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nj5jb" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.767086    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"]
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.768429    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.770582    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.770793    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.777659    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45769e7a-e7df-4d98-b779-aa349f8aa716-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.777716    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/45769e7a-e7df-4d98-b779-aa349f8aa716-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.784766    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"]
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.879154    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45769e7a-e7df-4d98-b779-aa349f8aa716-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.879241    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/45769e7a-e7df-4d98-b779-aa349f8aa716-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.879424    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/45769e7a-e7df-4d98-b779-aa349f8aa716-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:06 crc kubenswrapper[4824]: I0122 09:46:06.900434    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45769e7a-e7df-4d98-b779-aa349f8aa716-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") " pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:07 crc kubenswrapper[4824]: I0122 09:46:07.134885    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:08 crc kubenswrapper[4824]: E0122 09:46:08.136028    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qvd2r" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab"
Jan 22 09:46:08 crc kubenswrapper[4824]: E0122 09:46:08.136253    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nj5jb" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31"
Jan 22 09:46:11 crc kubenswrapper[4824]: I0122 09:46:11.966716    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"]
Jan 22 09:46:11 crc kubenswrapper[4824]: I0122 09:46:11.967554    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:11 crc kubenswrapper[4824]: I0122 09:46:11.979164    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"]
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.143095    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.143187    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kube-api-access\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.143212    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-var-lock\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.244643    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kube-api-access\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.244730    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-var-lock\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.244832    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.244967    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.245009    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-var-lock\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.265213    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kube-api-access\") pod \"installer-9-crc\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") " pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:12 crc kubenswrapper[4824]: I0122 09:46:12.299289    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:46:13 crc kubenswrapper[4824]: E0122 09:46:13.840519    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18"
Jan 22 09:46:13 crc kubenswrapper[4824]: E0122 09:46:13.840866    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgxdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xfnlv_openshift-marketplace(05d38c4c-ca18-478a-bcd5-01800ded772a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:13 crc kubenswrapper[4824]: E0122 09:46:13.842113    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xfnlv" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a"
Jan 22 09:46:15 crc kubenswrapper[4824]: E0122 09:46:15.893304    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18"
Jan 22 09:46:15 crc kubenswrapper[4824]: E0122 09:46:15.893640    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-88zfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bkmd8_openshift-marketplace(2b498256-7f0d-4941-998e-e670f5941cdb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:15 crc kubenswrapper[4824]: E0122 09:46:15.895449    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bkmd8" podUID="2b498256-7f0d-4941-998e-e670f5941cdb"
Jan 22 09:46:16 crc kubenswrapper[4824]: E0122 09:46:16.039032    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xfnlv" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a"
Jan 22 09:46:19 crc kubenswrapper[4824]: E0122 09:46:19.601836    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18"
Jan 22 09:46:19 crc kubenswrapper[4824]: E0122 09:46:19.602410    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8phnl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hv6zw_openshift-marketplace(cfe569b9-ef9e-4ce9-b985-511797cc0633): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:19 crc kubenswrapper[4824]: E0122 09:46:19.604215    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hv6zw" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.091881    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hv6zw" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.092129    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bkmd8" podUID="2b498256-7f0d-4941-998e-e670f5941cdb"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.175284    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.175686    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-trrnh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9wlq2_openshift-marketplace(2fd47eee-b6b0-4448-8bcd-09c188762714): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.177131    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9wlq2" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.197286    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.197491    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xdkr9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9rdhx_openshift-marketplace(072ef9f4-8754-45c5-9600-accb760d786c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.198655    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9rdhx" podUID="072ef9f4-8754-45c5-9600-accb760d786c"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.207371    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.207500    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9nbgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xfqfs_openshift-marketplace(8d2678ef-e863-4181-a497-b98b7fff0cce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.208837    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xfqfs" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce"
Jan 22 09:46:21 crc kubenswrapper[4824]: I0122 09:46:21.567274    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xxtjz" event={"ID":"f43b4618-291a-40d3-b1ca-32abfadaf376","Type":"ContainerStarted","Data":"6c81eb4037593614de4a1e4ab388e9bc564b2fa36de4a303ef0452f6d5b16d63"}
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.567872    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9rdhx" podUID="072ef9f4-8754-45c5-9600-accb760d786c"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.568569    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xfqfs" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce"
Jan 22 09:46:21 crc kubenswrapper[4824]: E0122 09:46:21.568758    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9wlq2" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714"
Jan 22 09:46:21 crc kubenswrapper[4824]: I0122 09:46:21.589708    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xxtjz" podStartSLOduration=189.589688677 podStartE2EDuration="3m9.589688677s" podCreationTimestamp="2026-01-22 09:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:21.582893693 +0000 UTC m=+214.486370686" watchObservedRunningTime="2026-01-22 09:46:21.589688677 +0000 UTC m=+214.493165660"
Jan 22 09:46:21 crc kubenswrapper[4824]: I0122 09:46:21.623862    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"]
Jan 22 09:46:21 crc kubenswrapper[4824]: I0122 09:46:21.643910    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"]
Jan 22 09:46:21 crc kubenswrapper[4824]: I0122 09:46:21.705852    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"]
Jan 22 09:46:21 crc kubenswrapper[4824]: W0122 09:46:21.714533    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1e4e9337_a9f7_42b2_b7a4_5cd46ae9b277.slice/crio-4acf91c7aad6ece55a8e1ac7ff7d965a0cef6a38db866b15a05da7faca6fe939 WatchSource:0}: Error finding container 4acf91c7aad6ece55a8e1ac7ff7d965a0cef6a38db866b15a05da7faca6fe939: Status 404 returned error can't find the container with id 4acf91c7aad6ece55a8e1ac7ff7d965a0cef6a38db866b15a05da7faca6fe939
Jan 22 09:46:21 crc kubenswrapper[4824]: I0122 09:46:21.720129    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"]
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.573215    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277","Type":"ContainerStarted","Data":"fe67e163ce0b11f92eb7cdc9f87893c7481769fce03b10ca080a554ee0df42c5"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.573542    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277","Type":"ContainerStarted","Data":"4acf91c7aad6ece55a8e1ac7ff7d965a0cef6a38db866b15a05da7faca6fe939"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.575066    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"45769e7a-e7df-4d98-b779-aa349f8aa716","Type":"ContainerStarted","Data":"7c8ff018166f879943fa3bb39886e0058d8234e4aa44f0fcdb2781cb57d9d9bf"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.575103    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"45769e7a-e7df-4d98-b779-aa349f8aa716","Type":"ContainerStarted","Data":"1fa2db34de9ec775cf5be6596b3342cbd7cfdcfa13dcb367df8e7f62e35fdd9d"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.577173    4824 generic.go:334] "Generic (PLEG): container finished" podID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerID="7742bb669ebb371304c069b943f41fb6f0310aba705fbfcb09c46448be25cfa2" exitCode=0
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.577212    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qvd2r" event={"ID":"88bd6885-a443-4d1f-b06e-5dd0f3107dab","Type":"ContainerDied","Data":"7742bb669ebb371304c069b943f41fb6f0310aba705fbfcb09c46448be25cfa2"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.578666    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls" event={"ID":"7ec14957-e115-4061-9db2-b1dce68f76ad","Type":"ContainerStarted","Data":"2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.578707    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls" event={"ID":"7ec14957-e115-4061-9db2-b1dce68f76ad","Type":"ContainerStarted","Data":"ecfaf56a8d79bce17276becfa53a929e771e8a8abb4ec34a415164e98d4cad9b"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.578743    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls" podUID="7ec14957-e115-4061-9db2-b1dce68f76ad" containerName="controller-manager" containerID="cri-o://2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869" gracePeriod=30
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.578971    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.583687    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" event={"ID":"44acb8f7-faae-48c5-9551-0e6bbb0d7afe","Type":"ContainerStarted","Data":"7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.583716    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" event={"ID":"44acb8f7-faae-48c5-9551-0e6bbb0d7afe","Type":"ContainerStarted","Data":"8b4fb24fe60d8f9881cdd0284161cdb9a72110d7b77a14ec0d351ada144952af"}
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.583828    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" podUID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" containerName="route-controller-manager" containerID="cri-o://7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7" gracePeriod=30
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.584103    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.588116    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.607298    4824 patch_prober.go:28] interesting pod/route-controller-manager-78f64778cd-v5lt4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": EOF" start-of-body=
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.607372    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" podUID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": EOF"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.621405    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=11.621387078 podStartE2EDuration="11.621387078s" podCreationTimestamp="2026-01-22 09:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:22.598926622 +0000 UTC m=+215.502403615" watchObservedRunningTime="2026-01-22 09:46:22.621387078 +0000 UTC m=+215.524864071"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.639013    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls" podStartSLOduration=42.63899254 podStartE2EDuration="42.63899254s" podCreationTimestamp="2026-01-22 09:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:22.622793614 +0000 UTC m=+215.526270607" watchObservedRunningTime="2026-01-22 09:46:22.63899254 +0000 UTC m=+215.542469533"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.639884    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" podStartSLOduration=42.639880683 podStartE2EDuration="42.639880683s" podCreationTimestamp="2026-01-22 09:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:22.639114773 +0000 UTC m=+215.542591766" watchObservedRunningTime="2026-01-22 09:46:22.639880683 +0000 UTC m=+215.543357676"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.656823    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=16.656802687 podStartE2EDuration="16.656802687s" podCreationTimestamp="2026-01-22 09:46:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:22.654516758 +0000 UTC m=+215.557993761" watchObservedRunningTime="2026-01-22 09:46:22.656802687 +0000 UTC m=+215.560279680"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.963160    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.974285    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.992059    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59865466c4-rg56g"]
Jan 22 09:46:22 crc kubenswrapper[4824]: E0122 09:46:22.992249    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" containerName="route-controller-manager"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.992260    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" containerName="route-controller-manager"
Jan 22 09:46:22 crc kubenswrapper[4824]: E0122 09:46:22.992271    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec14957-e115-4061-9db2-b1dce68f76ad" containerName="controller-manager"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.992276    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec14957-e115-4061-9db2-b1dce68f76ad" containerName="controller-manager"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.992399    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ec14957-e115-4061-9db2-b1dce68f76ad" containerName="controller-manager"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.992415    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" containerName="route-controller-manager"
Jan 22 09:46:22 crc kubenswrapper[4824]: I0122 09:46:22.992741    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.010603    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59865466c4-rg56g"]
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087766    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec14957-e115-4061-9db2-b1dce68f76ad-serving-cert\") pod \"7ec14957-e115-4061-9db2-b1dce68f76ad\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087803    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-client-ca\") pod \"7ec14957-e115-4061-9db2-b1dce68f76ad\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087847    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-client-ca\") pod \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087877    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bvtp\" (UniqueName: \"kubernetes.io/projected/7ec14957-e115-4061-9db2-b1dce68f76ad-kube-api-access-9bvtp\") pod \"7ec14957-e115-4061-9db2-b1dce68f76ad\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087897    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpz4s\" (UniqueName: \"kubernetes.io/projected/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-kube-api-access-lpz4s\") pod \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087916    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-config\") pod \"7ec14957-e115-4061-9db2-b1dce68f76ad\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.087978    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-proxy-ca-bundles\") pod \"7ec14957-e115-4061-9db2-b1dce68f76ad\" (UID: \"7ec14957-e115-4061-9db2-b1dce68f76ad\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088000    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-serving-cert\") pod \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088021    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-config\") pod \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\" (UID: \"44acb8f7-faae-48c5-9551-0e6bbb0d7afe\") "
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088193    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-proxy-ca-bundles\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088214    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-config\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088241    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb9sm\" (UniqueName: \"kubernetes.io/projected/83a4628b-6931-43ac-9baa-0440b1a079a1-kube-api-access-mb9sm\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088259    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-client-ca\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088281    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a4628b-6931-43ac-9baa-0440b1a079a1-serving-cert\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088736    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-client-ca" (OuterVolumeSpecName: "client-ca") pod "44acb8f7-faae-48c5-9551-0e6bbb0d7afe" (UID: "44acb8f7-faae-48c5-9551-0e6bbb0d7afe"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.088834    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7ec14957-e115-4061-9db2-b1dce68f76ad" (UID: "7ec14957-e115-4061-9db2-b1dce68f76ad"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.089086    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-client-ca" (OuterVolumeSpecName: "client-ca") pod "7ec14957-e115-4061-9db2-b1dce68f76ad" (UID: "7ec14957-e115-4061-9db2-b1dce68f76ad"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.089250    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-config" (OuterVolumeSpecName: "config") pod "7ec14957-e115-4061-9db2-b1dce68f76ad" (UID: "7ec14957-e115-4061-9db2-b1dce68f76ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.089565    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-config" (OuterVolumeSpecName: "config") pod "44acb8f7-faae-48c5-9551-0e6bbb0d7afe" (UID: "44acb8f7-faae-48c5-9551-0e6bbb0d7afe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.093322    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-kube-api-access-lpz4s" (OuterVolumeSpecName: "kube-api-access-lpz4s") pod "44acb8f7-faae-48c5-9551-0e6bbb0d7afe" (UID: "44acb8f7-faae-48c5-9551-0e6bbb0d7afe"). InnerVolumeSpecName "kube-api-access-lpz4s". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.094307    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ec14957-e115-4061-9db2-b1dce68f76ad-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7ec14957-e115-4061-9db2-b1dce68f76ad" (UID: "7ec14957-e115-4061-9db2-b1dce68f76ad"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.094364    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "44acb8f7-faae-48c5-9551-0e6bbb0d7afe" (UID: "44acb8f7-faae-48c5-9551-0e6bbb0d7afe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.094420    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ec14957-e115-4061-9db2-b1dce68f76ad-kube-api-access-9bvtp" (OuterVolumeSpecName: "kube-api-access-9bvtp") pod "7ec14957-e115-4061-9db2-b1dce68f76ad" (UID: "7ec14957-e115-4061-9db2-b1dce68f76ad"). InnerVolumeSpecName "kube-api-access-9bvtp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.189596    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-proxy-ca-bundles\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.189917    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-config\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.189954    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb9sm\" (UniqueName: \"kubernetes.io/projected/83a4628b-6931-43ac-9baa-0440b1a079a1-kube-api-access-mb9sm\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.189980    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-client-ca\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190005    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a4628b-6931-43ac-9baa-0440b1a079a1-serving-cert\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190049    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-proxy-ca-bundles\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190063    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190090    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190102    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ec14957-e115-4061-9db2-b1dce68f76ad-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190221    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190367    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190867    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bvtp\" (UniqueName: \"kubernetes.io/projected/7ec14957-e115-4061-9db2-b1dce68f76ad-kube-api-access-9bvtp\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190887    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpz4s\" (UniqueName: \"kubernetes.io/projected/44acb8f7-faae-48c5-9551-0e6bbb0d7afe-kube-api-access-lpz4s\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.190897    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec14957-e115-4061-9db2-b1dce68f76ad-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.191234    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-client-ca\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.191407    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-proxy-ca-bundles\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.191624    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-config\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.194109    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a4628b-6931-43ac-9baa-0440b1a079a1-serving-cert\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.207412    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb9sm\" (UniqueName: \"kubernetes.io/projected/83a4628b-6931-43ac-9baa-0440b1a079a1-kube-api-access-mb9sm\") pod \"controller-manager-59865466c4-rg56g\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") " pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.314137    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.591000    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.591015    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls" event={"ID":"7ec14957-e115-4061-9db2-b1dce68f76ad","Type":"ContainerDied","Data":"2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.590913    4824 generic.go:334] "Generic (PLEG): container finished" podID="7ec14957-e115-4061-9db2-b1dce68f76ad" containerID="2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869" exitCode=0
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.591408    4824 scope.go:117] "RemoveContainer" containerID="2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.591568    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls" event={"ID":"7ec14957-e115-4061-9db2-b1dce68f76ad","Type":"ContainerDied","Data":"ecfaf56a8d79bce17276becfa53a929e771e8a8abb4ec34a415164e98d4cad9b"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.601100    4824 generic.go:334] "Generic (PLEG): container finished" podID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" containerID="7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7" exitCode=0
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.601149    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" event={"ID":"44acb8f7-faae-48c5-9551-0e6bbb0d7afe","Type":"ContainerDied","Data":"7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.601156    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.601186    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4" event={"ID":"44acb8f7-faae-48c5-9551-0e6bbb0d7afe","Type":"ContainerDied","Data":"8b4fb24fe60d8f9881cdd0284161cdb9a72110d7b77a14ec0d351ada144952af"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.606762    4824 generic.go:334] "Generic (PLEG): container finished" podID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerID="3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0" exitCode=0
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.606907    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nj5jb" event={"ID":"1ddf148c-bafe-480f-b156-e3daa798bb31","Type":"ContainerDied","Data":"3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.608255    4824 generic.go:334] "Generic (PLEG): container finished" podID="45769e7a-e7df-4d98-b779-aa349f8aa716" containerID="7c8ff018166f879943fa3bb39886e0058d8234e4aa44f0fcdb2781cb57d9d9bf" exitCode=0
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.608428    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"]
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.608452    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"45769e7a-e7df-4d98-b779-aa349f8aa716","Type":"ContainerDied","Data":"7c8ff018166f879943fa3bb39886e0058d8234e4aa44f0fcdb2781cb57d9d9bf"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.611766    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5d9b8d9bd5-b6lls"]
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.611819    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qvd2r" event={"ID":"88bd6885-a443-4d1f-b06e-5dd0f3107dab","Type":"ContainerStarted","Data":"40981f624a90205622eab9f1ca631923b8deed71d17240a64a0f9f2432fa979b"}
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.618670    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"]
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.620598    4824 scope.go:117] "RemoveContainer" containerID="2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.623005    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78f64778cd-v5lt4"]
Jan 22 09:46:23 crc kubenswrapper[4824]: E0122 09:46:23.624785    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869\": container with ID starting with 2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869 not found: ID does not exist" containerID="2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.624825    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869"} err="failed to get container status \"2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869\": rpc error: code = NotFound desc = could not find container \"2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869\": container with ID starting with 2b125fbf74713f4a585d3e9ace6ad1341a7a58fd8af02a93dda41d03249de869 not found: ID does not exist"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.624902    4824 scope.go:117] "RemoveContainer" containerID="7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.644156    4824 scope.go:117] "RemoveContainer" containerID="7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7"
Jan 22 09:46:23 crc kubenswrapper[4824]: E0122 09:46:23.645992    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7\": container with ID starting with 7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7 not found: ID does not exist" containerID="7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.646122    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7"} err="failed to get container status \"7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7\": rpc error: code = NotFound desc = could not find container \"7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7\": container with ID starting with 7c85d62a82bcd8148819a89e141f7d8670704394d9322eec678fd15a15e7e8a7 not found: ID does not exist"
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.673283    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59865466c4-rg56g"]
Jan 22 09:46:23 crc kubenswrapper[4824]: I0122 09:46:23.677690    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qvd2r" podStartSLOduration=2.685204255 podStartE2EDuration="59.677668801s" podCreationTimestamp="2026-01-22 09:45:24 +0000 UTC" firstStartedPulling="2026-01-22 09:45:26.0089593 +0000 UTC m=+158.912436293" lastFinishedPulling="2026-01-22 09:46:23.001423846 +0000 UTC m=+215.904900839" observedRunningTime="2026-01-22 09:46:23.672007076 +0000 UTC m=+216.575484089" watchObservedRunningTime="2026-01-22 09:46:23.677668801 +0000 UTC m=+216.581145794"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.618255    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" event={"ID":"83a4628b-6931-43ac-9baa-0440b1a079a1","Type":"ContainerStarted","Data":"c1c00139df4c0a55ee5e143b4abd15e146bf039ef59b9b92190e0e520a8b6e60"}
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.618800    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" event={"ID":"83a4628b-6931-43ac-9baa-0440b1a079a1","Type":"ContainerStarted","Data":"87d162f489c833bae218defd8427675e51bd6d651fef808d1d9b99c3899a9725"}
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.618828    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.620521    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nj5jb" event={"ID":"1ddf148c-bafe-480f-b156-e3daa798bb31","Type":"ContainerStarted","Data":"a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf"}
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.624144    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.641091    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" podStartSLOduration=24.641068631 podStartE2EDuration="24.641068631s" podCreationTimestamp="2026-01-22 09:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:24.638847964 +0000 UTC m=+217.542324977" watchObservedRunningTime="2026-01-22 09:46:24.641068631 +0000 UTC m=+217.544545624"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.661046    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nj5jb" podStartSLOduration=3.197616731 podStartE2EDuration="1m0.661025943s" podCreationTimestamp="2026-01-22 09:45:24 +0000 UTC" firstStartedPulling="2026-01-22 09:45:27.014499814 +0000 UTC m=+159.917976807" lastFinishedPulling="2026-01-22 09:46:24.477909026 +0000 UTC m=+217.381386019" observedRunningTime="2026-01-22 09:46:24.659803982 +0000 UTC m=+217.563280995" watchObservedRunningTime="2026-01-22 09:46:24.661025943 +0000 UTC m=+217.564502936"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.716765    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.717124    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.798224    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:46:24 crc kubenswrapper[4824]: I0122 09:46:24.900945    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.022543    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/45769e7a-e7df-4d98-b779-aa349f8aa716-kubelet-dir\") pod \"45769e7a-e7df-4d98-b779-aa349f8aa716\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") "
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.022646    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45769e7a-e7df-4d98-b779-aa349f8aa716-kube-api-access\") pod \"45769e7a-e7df-4d98-b779-aa349f8aa716\" (UID: \"45769e7a-e7df-4d98-b779-aa349f8aa716\") "
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.022688    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45769e7a-e7df-4d98-b779-aa349f8aa716-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "45769e7a-e7df-4d98-b779-aa349f8aa716" (UID: "45769e7a-e7df-4d98-b779-aa349f8aa716"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.022908    4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/45769e7a-e7df-4d98-b779-aa349f8aa716-kubelet-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.043500    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45769e7a-e7df-4d98-b779-aa349f8aa716-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "45769e7a-e7df-4d98-b779-aa349f8aa716" (UID: "45769e7a-e7df-4d98-b779-aa349f8aa716"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.124185    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.124246    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.124680    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/45769e7a-e7df-4d98-b779-aa349f8aa716-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.410792    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44acb8f7-faae-48c5-9551-0e6bbb0d7afe" path="/var/lib/kubelet/pods/44acb8f7-faae-48c5-9551-0e6bbb0d7afe/volumes"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.411536    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ec14957-e115-4061-9db2-b1dce68f76ad" path="/var/lib/kubelet/pods/7ec14957-e115-4061-9db2-b1dce68f76ad/volumes"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.633868    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.634211    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"45769e7a-e7df-4d98-b779-aa349f8aa716","Type":"ContainerDied","Data":"1fa2db34de9ec775cf5be6596b3342cbd7cfdcfa13dcb367df8e7f62e35fdd9d"}
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.634233    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fa2db34de9ec775cf5be6596b3342cbd7cfdcfa13dcb367df8e7f62e35fdd9d"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.962763    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"]
Jan 22 09:46:25 crc kubenswrapper[4824]: E0122 09:46:25.963159    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45769e7a-e7df-4d98-b779-aa349f8aa716" containerName="pruner"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.963193    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="45769e7a-e7df-4d98-b779-aa349f8aa716" containerName="pruner"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.963465    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="45769e7a-e7df-4d98-b779-aa349f8aa716" containerName="pruner"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.964223    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.968590    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.972775    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"]
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.973375    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.973619    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.973826    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.973947    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt"
Jan 22 09:46:25 crc kubenswrapper[4824]: I0122 09:46:25.974021    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.139978    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-config\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.140296    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gj8t\" (UniqueName: \"kubernetes.io/projected/0d14a5b4-a963-4d01-9232-03163f1f3719-kube-api-access-6gj8t\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.140343    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d14a5b4-a963-4d01-9232-03163f1f3719-serving-cert\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.140422    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-client-ca\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.173465    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-nj5jb" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="registry-server" probeResult="failure" output=<
Jan 22 09:46:26 crc kubenswrapper[4824]:         timeout: failed to connect service ":50051" within 1s
Jan 22 09:46:26 crc kubenswrapper[4824]:  >
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.242031    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d14a5b4-a963-4d01-9232-03163f1f3719-serving-cert\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.242087    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-client-ca\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.242145    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-config\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.242163    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gj8t\" (UniqueName: \"kubernetes.io/projected/0d14a5b4-a963-4d01-9232-03163f1f3719-kube-api-access-6gj8t\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.243709    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-client-ca\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.243722    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-config\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.262773    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gj8t\" (UniqueName: \"kubernetes.io/projected/0d14a5b4-a963-4d01-9232-03163f1f3719-kube-api-access-6gj8t\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.262794    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d14a5b4-a963-4d01-9232-03163f1f3719-serving-cert\") pod \"route-controller-manager-769cc7cbb6-8s8qx\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") " pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.298915    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:26 crc kubenswrapper[4824]: I0122 09:46:26.699654    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"]
Jan 22 09:46:27 crc kubenswrapper[4824]: I0122 09:46:27.655005    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx" event={"ID":"0d14a5b4-a963-4d01-9232-03163f1f3719","Type":"ContainerStarted","Data":"a543023991d318190e4d47e5ed91798efb2630b234490b486de45a3ce2d62bed"}
Jan 22 09:46:27 crc kubenswrapper[4824]: I0122 09:46:27.655421    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx" event={"ID":"0d14a5b4-a963-4d01-9232-03163f1f3719","Type":"ContainerStarted","Data":"307f13849d20dcc53c64c4feb79953e5e043beda3aa4e9b708a6841330fbd902"}
Jan 22 09:46:27 crc kubenswrapper[4824]: I0122 09:46:27.655449    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:27 crc kubenswrapper[4824]: I0122 09:46:27.660055    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:27 crc kubenswrapper[4824]: I0122 09:46:27.670259    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx" podStartSLOduration=27.670235646 podStartE2EDuration="27.670235646s" podCreationTimestamp="2026-01-22 09:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:27.669836166 +0000 UTC m=+220.573313169" watchObservedRunningTime="2026-01-22 09:46:27.670235646 +0000 UTC m=+220.573712639"
Jan 22 09:46:30 crc kubenswrapper[4824]: I0122 09:46:30.333331    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:46:30 crc kubenswrapper[4824]: I0122 09:46:30.334287    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:46:30 crc kubenswrapper[4824]: I0122 09:46:30.334409    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:46:30 crc kubenswrapper[4824]: I0122 09:46:30.335019    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 09:46:30 crc kubenswrapper[4824]: I0122 09:46:30.335080    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784" gracePeriod=600
Jan 22 09:46:31 crc kubenswrapper[4824]: I0122 09:46:31.677627    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784" exitCode=0
Jan 22 09:46:31 crc kubenswrapper[4824]: I0122 09:46:31.677640    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784"}
Jan 22 09:46:34 crc kubenswrapper[4824]: I0122 09:46:34.767441    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:46:35 crc kubenswrapper[4824]: I0122 09:46:35.164408    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:46:35 crc kubenswrapper[4824]: I0122 09:46:35.207933    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:46:36 crc kubenswrapper[4824]: I0122 09:46:36.843726    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nj5jb"]
Jan 22 09:46:36 crc kubenswrapper[4824]: I0122 09:46:36.844537    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nj5jb" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="registry-server" containerID="cri-o://a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf" gracePeriod=2
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.375813    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.486789    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjbk6\" (UniqueName: \"kubernetes.io/projected/1ddf148c-bafe-480f-b156-e3daa798bb31-kube-api-access-rjbk6\") pod \"1ddf148c-bafe-480f-b156-e3daa798bb31\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") "
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.486854    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-catalog-content\") pod \"1ddf148c-bafe-480f-b156-e3daa798bb31\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") "
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.486878    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-utilities\") pod \"1ddf148c-bafe-480f-b156-e3daa798bb31\" (UID: \"1ddf148c-bafe-480f-b156-e3daa798bb31\") "
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.488730    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-utilities" (OuterVolumeSpecName: "utilities") pod "1ddf148c-bafe-480f-b156-e3daa798bb31" (UID: "1ddf148c-bafe-480f-b156-e3daa798bb31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.491992    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ddf148c-bafe-480f-b156-e3daa798bb31-kube-api-access-rjbk6" (OuterVolumeSpecName: "kube-api-access-rjbk6") pod "1ddf148c-bafe-480f-b156-e3daa798bb31" (UID: "1ddf148c-bafe-480f-b156-e3daa798bb31"). InnerVolumeSpecName "kube-api-access-rjbk6". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.549856    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ddf148c-bafe-480f-b156-e3daa798bb31" (UID: "1ddf148c-bafe-480f-b156-e3daa798bb31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.589761    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjbk6\" (UniqueName: \"kubernetes.io/projected/1ddf148c-bafe-480f-b156-e3daa798bb31-kube-api-access-rjbk6\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.589802    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.589816    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ddf148c-bafe-480f-b156-e3daa798bb31-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.708725    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"7b8ff90b63f9c4bc3e35eb99a32bf8aff84f6fad5953f3a33158e75564461092"}
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.711089    4824 generic.go:334] "Generic (PLEG): container finished" podID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerID="a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf" exitCode=0
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.711169    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nj5jb" event={"ID":"1ddf148c-bafe-480f-b156-e3daa798bb31","Type":"ContainerDied","Data":"a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf"}
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.711202    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nj5jb" event={"ID":"1ddf148c-bafe-480f-b156-e3daa798bb31","Type":"ContainerDied","Data":"46f5ab80d4eb8e567583e991512a1764cf52efd6f6abac59bae64941536406ce"}
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.711223    4824 scope.go:117] "RemoveContainer" containerID="a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.711381    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nj5jb"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.716158    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerStarted","Data":"03e91e93ac0e8925bf311e95107dd5dcad5fe1e95f212a41ef3a5937df6107c2"}
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.735642    4824 scope.go:117] "RemoveContainer" containerID="3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.760085    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nj5jb"]
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.762819    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nj5jb"]
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.770905    4824 scope.go:117] "RemoveContainer" containerID="891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.785317    4824 scope.go:117] "RemoveContainer" containerID="a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf"
Jan 22 09:46:37 crc kubenswrapper[4824]: E0122 09:46:37.785720    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf\": container with ID starting with a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf not found: ID does not exist" containerID="a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.785763    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf"} err="failed to get container status \"a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf\": rpc error: code = NotFound desc = could not find container \"a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf\": container with ID starting with a04922649a83e2ab6c955d500bc2a391236976e684aa01d98b92a2878df1fbdf not found: ID does not exist"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.785793    4824 scope.go:117] "RemoveContainer" containerID="3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0"
Jan 22 09:46:37 crc kubenswrapper[4824]: E0122 09:46:37.786172    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0\": container with ID starting with 3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0 not found: ID does not exist" containerID="3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.786203    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0"} err="failed to get container status \"3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0\": rpc error: code = NotFound desc = could not find container \"3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0\": container with ID starting with 3af0042293b848210a518dc021b63790a2e77b80414603ffc00b12bb63f8fde0 not found: ID does not exist"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.786243    4824 scope.go:117] "RemoveContainer" containerID="891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e"
Jan 22 09:46:37 crc kubenswrapper[4824]: E0122 09:46:37.786530    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e\": container with ID starting with 891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e not found: ID does not exist" containerID="891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e"
Jan 22 09:46:37 crc kubenswrapper[4824]: I0122 09:46:37.786556    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e"} err="failed to get container status \"891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e\": rpc error: code = NotFound desc = could not find container \"891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e\": container with ID starting with 891ed809195ba7e39c56f80f60082e502d6f4c949c130108bc7a024d8239368e not found: ID does not exist"
Jan 22 09:46:38 crc kubenswrapper[4824]: I0122 09:46:38.730016    4824 generic.go:334] "Generic (PLEG): container finished" podID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerID="03e91e93ac0e8925bf311e95107dd5dcad5fe1e95f212a41ef3a5937df6107c2" exitCode=0
Jan 22 09:46:38 crc kubenswrapper[4824]: I0122 09:46:38.730078    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerDied","Data":"03e91e93ac0e8925bf311e95107dd5dcad5fe1e95f212a41ef3a5937df6107c2"}
Jan 22 09:46:39 crc kubenswrapper[4824]: I0122 09:46:39.411171    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" path="/var/lib/kubelet/pods/1ddf148c-bafe-480f-b156-e3daa798bb31/volumes"
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.088390    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59865466c4-rg56g"]
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.089092    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" podUID="83a4628b-6931-43ac-9baa-0440b1a079a1" containerName="controller-manager" containerID="cri-o://c1c00139df4c0a55ee5e143b4abd15e146bf039ef59b9b92190e0e520a8b6e60" gracePeriod=30
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.180497    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"]
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.180744    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx" podUID="0d14a5b4-a963-4d01-9232-03163f1f3719" containerName="route-controller-manager" containerID="cri-o://a543023991d318190e4d47e5ed91798efb2630b234490b486de45a3ce2d62bed" gracePeriod=30
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.739835    4824 generic.go:334] "Generic (PLEG): container finished" podID="0d14a5b4-a963-4d01-9232-03163f1f3719" containerID="a543023991d318190e4d47e5ed91798efb2630b234490b486de45a3ce2d62bed" exitCode=0
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.739895    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx" event={"ID":"0d14a5b4-a963-4d01-9232-03163f1f3719","Type":"ContainerDied","Data":"a543023991d318190e4d47e5ed91798efb2630b234490b486de45a3ce2d62bed"}
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.741249    4824 generic.go:334] "Generic (PLEG): container finished" podID="83a4628b-6931-43ac-9baa-0440b1a079a1" containerID="c1c00139df4c0a55ee5e143b4abd15e146bf039ef59b9b92190e0e520a8b6e60" exitCode=0
Jan 22 09:46:40 crc kubenswrapper[4824]: I0122 09:46:40.741276    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" event={"ID":"83a4628b-6931-43ac-9baa-0440b1a079a1","Type":"ContainerDied","Data":"c1c00139df4c0a55ee5e143b4abd15e146bf039ef59b9b92190e0e520a8b6e60"}
Jan 22 09:46:43 crc kubenswrapper[4824]: I0122 09:46:43.315546    4824 patch_prober.go:28] interesting pod/controller-manager-59865466c4-rg56g container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body=
Jan 22 09:46:43 crc kubenswrapper[4824]: I0122 09:46:43.315642    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" podUID="83a4628b-6931-43ac-9baa-0440b1a079a1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.196806    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.226841    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9744fd7b7-svt7c"]
Jan 22 09:46:45 crc kubenswrapper[4824]: E0122 09:46:45.227029    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="extract-content"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227040    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="extract-content"
Jan 22 09:46:45 crc kubenswrapper[4824]: E0122 09:46:45.227056    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a4628b-6931-43ac-9baa-0440b1a079a1" containerName="controller-manager"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227063    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a4628b-6931-43ac-9baa-0440b1a079a1" containerName="controller-manager"
Jan 22 09:46:45 crc kubenswrapper[4824]: E0122 09:46:45.227195    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="registry-server"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227204    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="registry-server"
Jan 22 09:46:45 crc kubenswrapper[4824]: E0122 09:46:45.227211    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="extract-utilities"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227217    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="extract-utilities"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227311    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddf148c-bafe-480f-b156-e3daa798bb31" containerName="registry-server"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227322    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a4628b-6931-43ac-9baa-0440b1a079a1" containerName="controller-manager"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.227698    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.239665    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9744fd7b7-svt7c"]
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.288321    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a4628b-6931-43ac-9baa-0440b1a079a1-serving-cert\") pod \"83a4628b-6931-43ac-9baa-0440b1a079a1\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.288394    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-config\") pod \"83a4628b-6931-43ac-9baa-0440b1a079a1\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.288431    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-proxy-ca-bundles\") pod \"83a4628b-6931-43ac-9baa-0440b1a079a1\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.288479    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb9sm\" (UniqueName: \"kubernetes.io/projected/83a4628b-6931-43ac-9baa-0440b1a079a1-kube-api-access-mb9sm\") pod \"83a4628b-6931-43ac-9baa-0440b1a079a1\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.288515    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-client-ca\") pod \"83a4628b-6931-43ac-9baa-0440b1a079a1\" (UID: \"83a4628b-6931-43ac-9baa-0440b1a079a1\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.289295    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-client-ca" (OuterVolumeSpecName: "client-ca") pod "83a4628b-6931-43ac-9baa-0440b1a079a1" (UID: "83a4628b-6931-43ac-9baa-0440b1a079a1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.289322    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "83a4628b-6931-43ac-9baa-0440b1a079a1" (UID: "83a4628b-6931-43ac-9baa-0440b1a079a1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.289400    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-config" (OuterVolumeSpecName: "config") pod "83a4628b-6931-43ac-9baa-0440b1a079a1" (UID: "83a4628b-6931-43ac-9baa-0440b1a079a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.306542    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a4628b-6931-43ac-9baa-0440b1a079a1-kube-api-access-mb9sm" (OuterVolumeSpecName: "kube-api-access-mb9sm") pod "83a4628b-6931-43ac-9baa-0440b1a079a1" (UID: "83a4628b-6931-43ac-9baa-0440b1a079a1"). InnerVolumeSpecName "kube-api-access-mb9sm". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.306882    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a4628b-6931-43ac-9baa-0440b1a079a1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "83a4628b-6931-43ac-9baa-0440b1a079a1" (UID: "83a4628b-6931-43ac-9baa-0440b1a079a1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.341662    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389664    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4qp6\" (UniqueName: \"kubernetes.io/projected/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-kube-api-access-g4qp6\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389724    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-proxy-ca-bundles\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389753    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-client-ca\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389779    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-serving-cert\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389804    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-config\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389841    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389852    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a4628b-6931-43ac-9baa-0440b1a079a1-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389860    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389869    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a4628b-6931-43ac-9baa-0440b1a079a1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.389877    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb9sm\" (UniqueName: \"kubernetes.io/projected/83a4628b-6931-43ac-9baa-0440b1a079a1-kube-api-access-mb9sm\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.491150    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gj8t\" (UniqueName: \"kubernetes.io/projected/0d14a5b4-a963-4d01-9232-03163f1f3719-kube-api-access-6gj8t\") pod \"0d14a5b4-a963-4d01-9232-03163f1f3719\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.491347    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-config\") pod \"0d14a5b4-a963-4d01-9232-03163f1f3719\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.491422    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-client-ca\") pod \"0d14a5b4-a963-4d01-9232-03163f1f3719\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.491471    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d14a5b4-a963-4d01-9232-03163f1f3719-serving-cert\") pod \"0d14a5b4-a963-4d01-9232-03163f1f3719\" (UID: \"0d14a5b4-a963-4d01-9232-03163f1f3719\") "
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.491690    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-proxy-ca-bundles\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.491947    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-client-ca\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492106    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-client-ca" (OuterVolumeSpecName: "client-ca") pod "0d14a5b4-a963-4d01-9232-03163f1f3719" (UID: "0d14a5b4-a963-4d01-9232-03163f1f3719"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492117    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-serving-cert\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492175    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-config" (OuterVolumeSpecName: "config") pod "0d14a5b4-a963-4d01-9232-03163f1f3719" (UID: "0d14a5b4-a963-4d01-9232-03163f1f3719"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492271    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-config\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492409    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4qp6\" (UniqueName: \"kubernetes.io/projected/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-kube-api-access-g4qp6\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492531    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.492555    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d14a5b4-a963-4d01-9232-03163f1f3719-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.493065    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-proxy-ca-bundles\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.493647    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-client-ca\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.494386    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-config\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.495588    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d14a5b4-a963-4d01-9232-03163f1f3719-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0d14a5b4-a963-4d01-9232-03163f1f3719" (UID: "0d14a5b4-a963-4d01-9232-03163f1f3719"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.499275    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d14a5b4-a963-4d01-9232-03163f1f3719-kube-api-access-6gj8t" (OuterVolumeSpecName: "kube-api-access-6gj8t") pod "0d14a5b4-a963-4d01-9232-03163f1f3719" (UID: "0d14a5b4-a963-4d01-9232-03163f1f3719"). InnerVolumeSpecName "kube-api-access-6gj8t". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.499974    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-serving-cert\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.508169    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4qp6\" (UniqueName: \"kubernetes.io/projected/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-kube-api-access-g4qp6\") pod \"controller-manager-9744fd7b7-svt7c\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") " pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.551246    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.594144    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d14a5b4-a963-4d01-9232-03163f1f3719-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.594420    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gj8t\" (UniqueName: \"kubernetes.io/projected/0d14a5b4-a963-4d01-9232-03163f1f3719-kube-api-access-6gj8t\") on node \"crc\" DevicePath \"\""
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.769051    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.769021    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59865466c4-rg56g" event={"ID":"83a4628b-6931-43ac-9baa-0440b1a079a1","Type":"ContainerDied","Data":"87d162f489c833bae218defd8427675e51bd6d651fef808d1d9b99c3899a9725"}
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.769122    4824 scope.go:117] "RemoveContainer" containerID="c1c00139df4c0a55ee5e143b4abd15e146bf039ef59b9b92190e0e520a8b6e60"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.777024    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerStarted","Data":"6e4723e09d2c90df1133040b661458e056a5c4c1f00d93bf75e8c99376670084"}
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.778816    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx" event={"ID":"0d14a5b4-a963-4d01-9232-03163f1f3719","Type":"ContainerDied","Data":"307f13849d20dcc53c64c4feb79953e5e043beda3aa4e9b708a6841330fbd902"}
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.778833    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.788474    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerStarted","Data":"b97850d09dc3112ae65bb9eddd3293cd1e260d40f07ef8d03d8f610789ca85bb"}
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.789115    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59865466c4-rg56g"]
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.808143    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-59865466c4-rg56g"]
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.811947    4824 scope.go:117] "RemoveContainer" containerID="a543023991d318190e4d47e5ed91798efb2630b234490b486de45a3ce2d62bed"
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.831599    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"]
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.834035    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769cc7cbb6-8s8qx"]
Jan 22 09:46:45 crc kubenswrapper[4824]: I0122 09:46:45.883003    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9744fd7b7-svt7c"]
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.794923    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" event={"ID":"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38","Type":"ContainerStarted","Data":"b14917152ca3db5ef77f6d98b07a6756a25b338ce81ff5939fc5663198f7fc53"}
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.797162    4824 generic.go:334] "Generic (PLEG): container finished" podID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerID="4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9" exitCode=0
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.797245    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hv6zw" event={"ID":"cfe569b9-ef9e-4ce9-b985-511797cc0633","Type":"ContainerDied","Data":"4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9"}
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.802240    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerStarted","Data":"a3a6c2c04f313c45a5892849d9c800f01ecc2e84447a62a45ae85fce1d194ed4"}
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.804463    4824 generic.go:334] "Generic (PLEG): container finished" podID="072ef9f4-8754-45c5-9600-accb760d786c" containerID="7317b7210034e3aa2ec35c18a34feca5f748280edc9c30e6422e019e9b037d9f" exitCode=0
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.804516    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rdhx" event={"ID":"072ef9f4-8754-45c5-9600-accb760d786c","Type":"ContainerDied","Data":"7317b7210034e3aa2ec35c18a34feca5f748280edc9c30e6422e019e9b037d9f"}
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.806388    4824 generic.go:334] "Generic (PLEG): container finished" podID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerID="6e4723e09d2c90df1133040b661458e056a5c4c1f00d93bf75e8c99376670084" exitCode=0
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.806432    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerDied","Data":"6e4723e09d2c90df1133040b661458e056a5c4c1f00d93bf75e8c99376670084"}
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.809437    4824 generic.go:334] "Generic (PLEG): container finished" podID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerID="307754dde20943c0e432815035492fac2217bcd95c9dbd5f59d30bf021e0baba" exitCode=0
Jan 22 09:46:46 crc kubenswrapper[4824]: I0122 09:46:46.809839    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfqfs" event={"ID":"8d2678ef-e863-4181-a497-b98b7fff0cce","Type":"ContainerDied","Data":"307754dde20943c0e432815035492fac2217bcd95c9dbd5f59d30bf021e0baba"}
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.410696    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d14a5b4-a963-4d01-9232-03163f1f3719" path="/var/lib/kubelet/pods/0d14a5b4-a963-4d01-9232-03163f1f3719/volumes"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.411225    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a4628b-6931-43ac-9baa-0440b1a079a1" path="/var/lib/kubelet/pods/83a4628b-6931-43ac-9baa-0440b1a079a1/volumes"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.434579    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xfnlv" podStartSLOduration=4.212686741 podStartE2EDuration="1m19.43454788s" podCreationTimestamp="2026-01-22 09:45:28 +0000 UTC" firstStartedPulling="2026-01-22 09:45:30.094289607 +0000 UTC m=+162.997766600" lastFinishedPulling="2026-01-22 09:46:45.316150746 +0000 UTC m=+238.219627739" observedRunningTime="2026-01-22 09:46:46.93939704 +0000 UTC m=+239.842874113" watchObservedRunningTime="2026-01-22 09:46:47.43454788 +0000 UTC m=+240.338024873"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.814272    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" event={"ID":"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38","Type":"ContainerStarted","Data":"ec8e75cf7df9b51326581fdd0a2eaea619efb86432955ae4d521989852d32634"}
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.814759    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.816345    4824 generic.go:334] "Generic (PLEG): container finished" podID="2b498256-7f0d-4941-998e-e670f5941cdb" containerID="b97850d09dc3112ae65bb9eddd3293cd1e260d40f07ef8d03d8f610789ca85bb" exitCode=0
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.816409    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerDied","Data":"b97850d09dc3112ae65bb9eddd3293cd1e260d40f07ef8d03d8f610789ca85bb"}
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.818641    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.863038    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" podStartSLOduration=7.863012609 podStartE2EDuration="7.863012609s" podCreationTimestamp="2026-01-22 09:46:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:47.839467205 +0000 UTC m=+240.742944198" watchObservedRunningTime="2026-01-22 09:46:47.863012609 +0000 UTC m=+240.766489622"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.969015    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"]
Jan 22 09:46:47 crc kubenswrapper[4824]: E0122 09:46:47.969211    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d14a5b4-a963-4d01-9232-03163f1f3719" containerName="route-controller-manager"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.969222    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d14a5b4-a963-4d01-9232-03163f1f3719" containerName="route-controller-manager"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.969328    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d14a5b4-a963-4d01-9232-03163f1f3719" containerName="route-controller-manager"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.969728    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.975593    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.975614    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.976067    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.976097    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.976225    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.976310    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert"
Jan 22 09:46:47 crc kubenswrapper[4824]: I0122 09:46:47.982633    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"]
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.134879    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-client-ca\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.134946    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-config\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.135016    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85cs9\" (UniqueName: \"kubernetes.io/projected/3860e075-2372-4e81-aece-e4bb5ff66e7e-kube-api-access-85cs9\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.135090    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3860e075-2372-4e81-aece-e4bb5ff66e7e-serving-cert\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.236462    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-client-ca\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.236730    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-config\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.236955    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85cs9\" (UniqueName: \"kubernetes.io/projected/3860e075-2372-4e81-aece-e4bb5ff66e7e-kube-api-access-85cs9\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.237106    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3860e075-2372-4e81-aece-e4bb5ff66e7e-serving-cert\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.238396    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-client-ca\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.239841    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-config\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.242787    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3860e075-2372-4e81-aece-e4bb5ff66e7e-serving-cert\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.260557    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85cs9\" (UniqueName: \"kubernetes.io/projected/3860e075-2372-4e81-aece-e4bb5ff66e7e-kube-api-access-85cs9\") pod \"route-controller-manager-77f58dc8c-h4mhx\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") " pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.289113    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.524470    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.525115    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.791541    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"]
Jan 22 09:46:48 crc kubenswrapper[4824]: W0122 09:46:48.814449    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3860e075_2372_4e81_aece_e4bb5ff66e7e.slice/crio-ad5fccae04fe60700d46f3c1666c251e8c3ed3ff0326aab0417595a0b8cdd66d WatchSource:0}: Error finding container ad5fccae04fe60700d46f3c1666c251e8c3ed3ff0326aab0417595a0b8cdd66d: Status 404 returned error can't find the container with id ad5fccae04fe60700d46f3c1666c251e8c3ed3ff0326aab0417595a0b8cdd66d
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.825159    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rdhx" event={"ID":"072ef9f4-8754-45c5-9600-accb760d786c","Type":"ContainerStarted","Data":"1419fdfd5269cfff5e11ab0c006f882559be5f169910c27a72809be04a43be39"}
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.827036    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" event={"ID":"3860e075-2372-4e81-aece-e4bb5ff66e7e","Type":"ContainerStarted","Data":"ad5fccae04fe60700d46f3c1666c251e8c3ed3ff0326aab0417595a0b8cdd66d"}
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.829269    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hv6zw" event={"ID":"cfe569b9-ef9e-4ce9-b985-511797cc0633","Type":"ContainerStarted","Data":"991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a"}
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.851194    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9rdhx" podStartSLOduration=3.7095808999999997 podStartE2EDuration="1m24.851166555s" podCreationTimestamp="2026-01-22 09:45:24 +0000 UTC" firstStartedPulling="2026-01-22 09:45:27.017669853 +0000 UTC m=+159.921146846" lastFinishedPulling="2026-01-22 09:46:48.159255508 +0000 UTC m=+241.062732501" observedRunningTime="2026-01-22 09:46:48.843795576 +0000 UTC m=+241.747272569" watchObservedRunningTime="2026-01-22 09:46:48.851166555 +0000 UTC m=+241.754643548"
Jan 22 09:46:48 crc kubenswrapper[4824]: I0122 09:46:48.863114    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hv6zw" podStartSLOduration=3.567512421 podStartE2EDuration="1m22.86308427s" podCreationTimestamp="2026-01-22 09:45:26 +0000 UTC" firstStartedPulling="2026-01-22 09:45:29.045127333 +0000 UTC m=+161.948604326" lastFinishedPulling="2026-01-22 09:46:48.340699192 +0000 UTC m=+241.244176175" observedRunningTime="2026-01-22 09:46:48.859286193 +0000 UTC m=+241.762763186" watchObservedRunningTime="2026-01-22 09:46:48.86308427 +0000 UTC m=+241.766561263"
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.562283    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xfnlv" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="registry-server" probeResult="failure" output=<
Jan 22 09:46:49 crc kubenswrapper[4824]:         timeout: failed to connect service ":50051" within 1s
Jan 22 09:46:49 crc kubenswrapper[4824]:  >
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.835730    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" event={"ID":"3860e075-2372-4e81-aece-e4bb5ff66e7e","Type":"ContainerStarted","Data":"69fc867d077f92fdd2db80cceb03a2e106a1ea50be845d7d9442fb248de05920"}
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.835986    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.838529    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerStarted","Data":"a4c02e3b803ee63ea2c2298af19e018d640051c4f1cf7b2bee77291fc7b6c940"}
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.840506    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerStarted","Data":"02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb"}
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.842961    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfqfs" event={"ID":"8d2678ef-e863-4181-a497-b98b7fff0cce","Type":"ContainerStarted","Data":"2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921"}
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.843579    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.925872    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" podStartSLOduration=9.925848559 podStartE2EDuration="9.925848559s" podCreationTimestamp="2026-01-22 09:46:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:46:49.872577603 +0000 UTC m=+242.776054596" watchObservedRunningTime="2026-01-22 09:46:49.925848559 +0000 UTC m=+242.829325552"
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.926413    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xfqfs" podStartSLOduration=4.222663283 podStartE2EDuration="1m23.926407504s" podCreationTimestamp="2026-01-22 09:45:26 +0000 UTC" firstStartedPulling="2026-01-22 09:45:29.054242324 +0000 UTC m=+161.957719317" lastFinishedPulling="2026-01-22 09:46:48.757986545 +0000 UTC m=+241.661463538" observedRunningTime="2026-01-22 09:46:49.924178937 +0000 UTC m=+242.827655940" watchObservedRunningTime="2026-01-22 09:46:49.926407504 +0000 UTC m=+242.829884497"
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.972926    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bkmd8" podStartSLOduration=3.07516867 podStartE2EDuration="1m22.972905206s" podCreationTimestamp="2026-01-22 09:45:27 +0000 UTC" firstStartedPulling="2026-01-22 09:45:29.057767445 +0000 UTC m=+161.961244438" lastFinishedPulling="2026-01-22 09:46:48.955503981 +0000 UTC m=+241.858980974" observedRunningTime="2026-01-22 09:46:49.969480858 +0000 UTC m=+242.872957861" watchObservedRunningTime="2026-01-22 09:46:49.972905206 +0000 UTC m=+242.876382199"
Jan 22 09:46:49 crc kubenswrapper[4824]: I0122 09:46:49.993922    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wlq2" podStartSLOduration=4.31235136 podStartE2EDuration="1m25.993901655s" podCreationTimestamp="2026-01-22 09:45:24 +0000 UTC" firstStartedPulling="2026-01-22 09:45:27.021053719 +0000 UTC m=+159.924530712" lastFinishedPulling="2026-01-22 09:46:48.702603994 +0000 UTC m=+241.606081007" observedRunningTime="2026-01-22 09:46:49.993029443 +0000 UTC m=+242.896506436" watchObservedRunningTime="2026-01-22 09:46:49.993901655 +0000 UTC m=+242.897378648"
Jan 22 09:46:54 crc kubenswrapper[4824]: I0122 09:46:54.979308    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:46:54 crc kubenswrapper[4824]: I0122 09:46:54.979941    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:46:55 crc kubenswrapper[4824]: I0122 09:46:55.024084    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:46:55 crc kubenswrapper[4824]: I0122 09:46:55.296593    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:46:55 crc kubenswrapper[4824]: I0122 09:46:55.296657    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:46:55 crc kubenswrapper[4824]: I0122 09:46:55.341196    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:46:55 crc kubenswrapper[4824]: I0122 09:46:55.932104    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:46:55 crc kubenswrapper[4824]: I0122 09:46:55.939240    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:46:56 crc kubenswrapper[4824]: I0122 09:46:56.661508    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:46:56 crc kubenswrapper[4824]: I0122 09:46:56.661560    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:46:56 crc kubenswrapper[4824]: I0122 09:46:56.721366    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:46:56 crc kubenswrapper[4824]: I0122 09:46:56.933615    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:46:57 crc kubenswrapper[4824]: I0122 09:46:57.069788    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:46:57 crc kubenswrapper[4824]: I0122 09:46:57.069853    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:46:57 crc kubenswrapper[4824]: I0122 09:46:57.117471    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:46:57 crc kubenswrapper[4824]: I0122 09:46:57.437934    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wlq2"]
Jan 22 09:46:57 crc kubenswrapper[4824]: I0122 09:46:57.893340    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wlq2" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="registry-server" containerID="cri-o://02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb" gracePeriod=2
Jan 22 09:46:57 crc kubenswrapper[4824]: I0122 09:46:57.933808    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:46:58 crc kubenswrapper[4824]: I0122 09:46:58.087672    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:46:58 crc kubenswrapper[4824]: I0122 09:46:58.088032    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:46:58 crc kubenswrapper[4824]: I0122 09:46:58.124606    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:46:58 crc kubenswrapper[4824]: I0122 09:46:58.574517    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:46:58 crc kubenswrapper[4824]: I0122 09:46:58.615867    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:46:58 crc kubenswrapper[4824]: I0122 09:46:58.961284    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:46:59 crc kubenswrapper[4824]: I0122 09:46:59.834986    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hv6zw"]
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.071213    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hv6zw" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="registry-server" containerID="cri-o://991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a" gracePeriod=2
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.121534    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9744fd7b7-svt7c"]
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.121800    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" containerName="controller-manager" containerID="cri-o://ec8e75cf7df9b51326581fdd0a2eaea619efb86432955ae4d521989852d32634" gracePeriod=30
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.143993    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"]
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.144648    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" containerName="route-controller-manager" containerID="cri-o://69fc867d077f92fdd2db80cceb03a2e106a1ea50be845d7d9442fb248de05920" gracePeriod=30
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.153607    4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"]
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.153881    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970" gracePeriod=15
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.153998    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12" gracePeriod=15
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.154014    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f" gracePeriod=15
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.154005    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4" gracePeriod=15
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.154079    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335" gracePeriod=15
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156011    4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"]
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156222    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156232    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints"
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156245    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156251    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver"
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156257    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156262    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer"
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156272    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156278    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz"
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156290    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156295    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller"
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156306    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156311    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156407    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156417    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156426    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156435    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156441    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller"
Jan 22 09:47:00 crc kubenswrapper[4824]: E0122 09:47:00.156526    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156532    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.156622    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.159052    4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"]
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.159840    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.181746    4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.213872    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"]
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262173    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262231    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262267    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262293    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262331    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262368    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262388    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.262411    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363515    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363584    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363609    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363648    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363676    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363722    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363749    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363770    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363849    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363933    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363967    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.363998    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.364025    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.364054    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.364084    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.364113    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: I0122 09:47:00.527669    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:00 crc kubenswrapper[4824]: W0122 09:47:00.544593    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-a63f686f8847c1c9a6a55523f9d9318235c0c07d0eb87a8ba0b3b0ebe3649df6 WatchSource:0}: Error finding container a63f686f8847c1c9a6a55523f9d9318235c0c07d0eb87a8ba0b3b0ebe3649df6: Status 404 returned error can't find the container with id a63f686f8847c1c9a6a55523f9d9318235c0c07d0eb87a8ba0b3b0ebe3649df6
Jan 22 09:47:01 crc kubenswrapper[4824]: I0122 09:47:01.076463    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a63f686f8847c1c9a6a55523f9d9318235c0c07d0eb87a8ba0b3b0ebe3649df6"}
Jan 22 09:47:03 crc kubenswrapper[4824]: I0122 09:47:03.092724    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log"
Jan 22 09:47:03 crc kubenswrapper[4824]: I0122 09:47:03.094233    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log"
Jan 22 09:47:03 crc kubenswrapper[4824]: I0122 09:47:03.095052    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f" exitCode=2
Jan 22 09:47:04 crc kubenswrapper[4824]: E0122 09:47:04.435756    4824 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" volumeName="registry-storage"
Jan 22 09:47:05 crc kubenswrapper[4824]: E0122 09:47:05.297482    4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb is running failed: container process not found" containerID="02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb" cmd=["grpc_health_probe","-addr=:50051"]
Jan 22 09:47:05 crc kubenswrapper[4824]: E0122 09:47:05.298282    4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb is running failed: container process not found" containerID="02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb" cmd=["grpc_health_probe","-addr=:50051"]
Jan 22 09:47:05 crc kubenswrapper[4824]: E0122 09:47:05.298937    4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb is running failed: container process not found" containerID="02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb" cmd=["grpc_health_probe","-addr=:50051"]
Jan 22 09:47:05 crc kubenswrapper[4824]: E0122 09:47:05.298990    4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-9wlq2" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="registry-server"
Jan 22 09:47:05 crc kubenswrapper[4824]: E0122 09:47:05.299232    4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-9wlq2.188d048b60e47149  openshift-marketplace    0 0001-01-01 00:00:00 +0000 UTC   map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-9wlq2,UID:2fd47eee-b6b0-4448-8bcd-09c188762714,APIVersion:v1,ResourceVersion:28328,FieldPath:spec.containers{registry-server},},Reason:Unhealthy,Message:Readiness probe errored: rpc error: code = NotFound desc = container is not created or running: checking if PID of 02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb is running failed: container process not found,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:47:05.299022153 +0000 UTC m=+258.202499146,LastTimestamp:2026-01-22 09:47:05.299022153 +0000 UTC m=+258.202499146,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}"
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.552374    4824 patch_prober.go:28] interesting pod/controller-manager-9744fd7b7-svt7c container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body=
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.552956    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused"
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.933316    4824 generic.go:334] "Generic (PLEG): container finished" podID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" containerID="ec8e75cf7df9b51326581fdd0a2eaea619efb86432955ae4d521989852d32634" exitCode=0
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.933412    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" event={"ID":"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38","Type":"ContainerDied","Data":"ec8e75cf7df9b51326581fdd0a2eaea619efb86432955ae4d521989852d32634"}
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.935506    4824 generic.go:334] "Generic (PLEG): container finished" podID="3860e075-2372-4e81-aece-e4bb5ff66e7e" containerID="69fc867d077f92fdd2db80cceb03a2e106a1ea50be845d7d9442fb248de05920" exitCode=0
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.935577    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" event={"ID":"3860e075-2372-4e81-aece-e4bb5ff66e7e","Type":"ContainerDied","Data":"69fc867d077f92fdd2db80cceb03a2e106a1ea50be845d7d9442fb248de05920"}
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.937597    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log"
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.939937    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log"
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.940630    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12" exitCode=0
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.940649    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335" exitCode=0
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.940656    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4" exitCode=0
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.940716    4824 scope.go:117] "RemoveContainer" containerID="86219c1efe553644f43ae0ea554de78a66cdbbb96917e7ea665bdc034e00f5fe"
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.943418    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9wlq2_2fd47eee-b6b0-4448-8bcd-09c188762714/registry-server/0.log"
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.944266    4824 generic.go:334] "Generic (PLEG): container finished" podID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerID="02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb" exitCode=137
Jan 22 09:47:05 crc kubenswrapper[4824]: I0122 09:47:05.944299    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerDied","Data":"02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.075260    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9wlq2_2fd47eee-b6b0-4448-8bcd-09c188762714/registry-server/0.log"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.076256    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.077066    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.077271    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: E0122 09:47:06.079812    4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-9wlq2.188d048b60e47149  openshift-marketplace    0 0001-01-01 00:00:00 +0000 UTC   map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-9wlq2,UID:2fd47eee-b6b0-4448-8bcd-09c188762714,APIVersion:v1,ResourceVersion:28328,FieldPath:spec.containers{registry-server},},Reason:Unhealthy,Message:Readiness probe errored: rpc error: code = NotFound desc = container is not created or running: checking if PID of 02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb is running failed: container process not found,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-22 09:47:05.299022153 +0000 UTC m=+258.202499146,LastTimestamp:2026-01-22 09:47:05.299022153 +0000 UTC m=+258.202499146,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.137251    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trrnh\" (UniqueName: \"kubernetes.io/projected/2fd47eee-b6b0-4448-8bcd-09c188762714-kube-api-access-trrnh\") pod \"2fd47eee-b6b0-4448-8bcd-09c188762714\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.137391    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-utilities\") pod \"2fd47eee-b6b0-4448-8bcd-09c188762714\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.137445    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-catalog-content\") pod \"2fd47eee-b6b0-4448-8bcd-09c188762714\" (UID: \"2fd47eee-b6b0-4448-8bcd-09c188762714\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.139669    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-utilities" (OuterVolumeSpecName: "utilities") pod "2fd47eee-b6b0-4448-8bcd-09c188762714" (UID: "2fd47eee-b6b0-4448-8bcd-09c188762714"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.143175    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd47eee-b6b0-4448-8bcd-09c188762714-kube-api-access-trrnh" (OuterVolumeSpecName: "kube-api-access-trrnh") pod "2fd47eee-b6b0-4448-8bcd-09c188762714" (UID: "2fd47eee-b6b0-4448-8bcd-09c188762714"). InnerVolumeSpecName "kube-api-access-trrnh". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.158558    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.159326    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.159608    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.159898    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.199114    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fd47eee-b6b0-4448-8bcd-09c188762714" (UID: "2fd47eee-b6b0-4448-8bcd-09c188762714"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238187    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-config\") pod \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238242    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-proxy-ca-bundles\") pod \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238282    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-client-ca\") pod \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238323    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4qp6\" (UniqueName: \"kubernetes.io/projected/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-kube-api-access-g4qp6\") pod \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238385    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-serving-cert\") pod \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\" (UID: \"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238758    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238785    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fd47eee-b6b0-4448-8bcd-09c188762714-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238799    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trrnh\" (UniqueName: \"kubernetes.io/projected/2fd47eee-b6b0-4448-8bcd-09c188762714-kube-api-access-trrnh\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238942    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" (UID: "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.238954    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-client-ca" (OuterVolumeSpecName: "client-ca") pod "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" (UID: "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.239032    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-config" (OuterVolumeSpecName: "config") pod "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" (UID: "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.242020    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" (UID: "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.242126    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-kube-api-access-g4qp6" (OuterVolumeSpecName: "kube-api-access-g4qp6") pod "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" (UID: "4ce653fe-67cf-4a94-bff2-cf4ef16b4e38"). InnerVolumeSpecName "kube-api-access-g4qp6". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.340409    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.340651    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4qp6\" (UniqueName: \"kubernetes.io/projected/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-kube-api-access-g4qp6\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.340661    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.340672    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.340680    4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38-proxy-ca-bundles\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.539750    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.543681    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.548007    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.548607    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.549040    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.549393    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647482    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647523    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647630    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647623    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647646    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647668    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647866    4824 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647877    4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.647885    4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.655393    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.655860    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.656022    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.656176    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.656327    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.656525    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.748248    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3860e075-2372-4e81-aece-e4bb5ff66e7e-serving-cert\") pod \"3860e075-2372-4e81-aece-e4bb5ff66e7e\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.748637    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85cs9\" (UniqueName: \"kubernetes.io/projected/3860e075-2372-4e81-aece-e4bb5ff66e7e-kube-api-access-85cs9\") pod \"3860e075-2372-4e81-aece-e4bb5ff66e7e\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.748685    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-config\") pod \"3860e075-2372-4e81-aece-e4bb5ff66e7e\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.748732    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-client-ca\") pod \"3860e075-2372-4e81-aece-e4bb5ff66e7e\" (UID: \"3860e075-2372-4e81-aece-e4bb5ff66e7e\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.749504    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-client-ca" (OuterVolumeSpecName: "client-ca") pod "3860e075-2372-4e81-aece-e4bb5ff66e7e" (UID: "3860e075-2372-4e81-aece-e4bb5ff66e7e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.749518    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-config" (OuterVolumeSpecName: "config") pod "3860e075-2372-4e81-aece-e4bb5ff66e7e" (UID: "3860e075-2372-4e81-aece-e4bb5ff66e7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.754145    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.754214    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3860e075-2372-4e81-aece-e4bb5ff66e7e-kube-api-access-85cs9" (OuterVolumeSpecName: "kube-api-access-85cs9") pod "3860e075-2372-4e81-aece-e4bb5ff66e7e" (UID: "3860e075-2372-4e81-aece-e4bb5ff66e7e"). InnerVolumeSpecName "kube-api-access-85cs9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.754596    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3860e075-2372-4e81-aece-e4bb5ff66e7e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3860e075-2372-4e81-aece-e4bb5ff66e7e" (UID: "3860e075-2372-4e81-aece-e4bb5ff66e7e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.755617    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.756056    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.759941    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.763562    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.763850    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.764074    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850348    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-catalog-content\") pod \"cfe569b9-ef9e-4ce9-b985-511797cc0633\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850503    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8phnl\" (UniqueName: \"kubernetes.io/projected/cfe569b9-ef9e-4ce9-b985-511797cc0633-kube-api-access-8phnl\") pod \"cfe569b9-ef9e-4ce9-b985-511797cc0633\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850567    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-utilities\") pod \"cfe569b9-ef9e-4ce9-b985-511797cc0633\" (UID: \"cfe569b9-ef9e-4ce9-b985-511797cc0633\") "
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850792    4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3860e075-2372-4e81-aece-e4bb5ff66e7e-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850811    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85cs9\" (UniqueName: \"kubernetes.io/projected/3860e075-2372-4e81-aece-e4bb5ff66e7e-kube-api-access-85cs9\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850822    4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.850831    4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3860e075-2372-4e81-aece-e4bb5ff66e7e-client-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.851651    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-utilities" (OuterVolumeSpecName: "utilities") pod "cfe569b9-ef9e-4ce9-b985-511797cc0633" (UID: "cfe569b9-ef9e-4ce9-b985-511797cc0633"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.853137    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe569b9-ef9e-4ce9-b985-511797cc0633-kube-api-access-8phnl" (OuterVolumeSpecName: "kube-api-access-8phnl") pod "cfe569b9-ef9e-4ce9-b985-511797cc0633" (UID: "cfe569b9-ef9e-4ce9-b985-511797cc0633"). InnerVolumeSpecName "kube-api-access-8phnl". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.879630    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cfe569b9-ef9e-4ce9-b985-511797cc0633" (UID: "cfe569b9-ef9e-4ce9-b985-511797cc0633"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951468    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951500    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8phnl\" (UniqueName: \"kubernetes.io/projected/cfe569b9-ef9e-4ce9-b985-511797cc0633-kube-api-access-8phnl\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951512    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cfe569b9-ef9e-4ce9-b985-511797cc0633-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951709    4824 generic.go:334] "Generic (PLEG): container finished" podID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerID="991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a" exitCode=0
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951767    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hv6zw" event={"ID":"cfe569b9-ef9e-4ce9-b985-511797cc0633","Type":"ContainerDied","Data":"991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951786    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hv6zw"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951797    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hv6zw" event={"ID":"cfe569b9-ef9e-4ce9-b985-511797cc0633","Type":"ContainerDied","Data":"af979d858887a89032987205e217b373121c24078b56810ce5bc4a6ff2fda3ec"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.951817    4824 scope.go:117] "RemoveContainer" containerID="991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.952449    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.953373    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.954230    4824 generic.go:334] "Generic (PLEG): container finished" podID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" containerID="fe67e163ce0b11f92eb7cdc9f87893c7481769fce03b10ca080a554ee0df42c5" exitCode=0
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.954300    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277","Type":"ContainerDied","Data":"fe67e163ce0b11f92eb7cdc9f87893c7481769fce03b10ca080a554ee0df42c5"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.954708    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.955449    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.955841    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9wlq2_2fd47eee-b6b0-4448-8bcd-09c188762714/registry-server/0.log"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.955912    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.956102    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.956527    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.956603    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wlq2" event={"ID":"2fd47eee-b6b0-4448-8bcd-09c188762714","Type":"ContainerDied","Data":"e106e71f1ac6b3af543619475b1236acef177c6ac8f4ee29cb9d38d8fb81440c"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.956640    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wlq2"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.957015    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.957409    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.957576    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.957927    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.958469    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.958720    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.958934    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.959089    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.959284    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.959411    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.959640    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.959868    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.960085    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.960333    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.960584    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.960739    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.960912    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.961193    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.961541    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.961653    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.961825    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.961827    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" event={"ID":"4ce653fe-67cf-4a94-bff2-cf4ef16b4e38","Type":"ContainerDied","Data":"b14917152ca3db5ef77f6d98b07a6756a25b338ce81ff5939fc5663198f7fc53"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.962014    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.962328    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.962623    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.962903    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.963235    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.963457    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.963801    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.964057    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.968054    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.968317    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" event={"ID":"3860e075-2372-4e81-aece-e4bb5ff66e7e","Type":"ContainerDied","Data":"ad5fccae04fe60700d46f3c1666c251e8c3ed3ff0326aab0417595a0b8cdd66d"}
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.968903    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.969217    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.969773    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.970254    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.970530    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.970739    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.972199    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.973577    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.973801    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.973969    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.974119    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.974261    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.974537    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.974694    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.981445    4824 scope.go:117] "RemoveContainer" containerID="4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.981711    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.982571    4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970" exitCode=0
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.982723    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.982720    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.983204    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.983481    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.984611    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.986222    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.988068    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.988337    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.989875    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.990230    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.990471    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.990667    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.990933    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.991153    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:06 crc kubenswrapper[4824]: I0122 09:47:06.991453    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.002611    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.002903    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.003377    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.003696    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.004075    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.004519    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.004851    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.006057    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.006629    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.006956    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.007218    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.007513    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.007883    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.008171    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.009020    4824 scope.go:117] "RemoveContainer" containerID="cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.059760    4824 scope.go:117] "RemoveContainer" containerID="991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.060199    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a\": container with ID starting with 991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a not found: ID does not exist" containerID="991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.060308    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a"} err="failed to get container status \"991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a\": rpc error: code = NotFound desc = could not find container \"991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a\": container with ID starting with 991072c5cf3489e48adc3f1a0f75e4170923e407b27f4f80a9c083bd04460a2a not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.060407    4824 scope.go:117] "RemoveContainer" containerID="4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.060895    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9\": container with ID starting with 4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9 not found: ID does not exist" containerID="4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.060950    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9"} err="failed to get container status \"4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9\": rpc error: code = NotFound desc = could not find container \"4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9\": container with ID starting with 4d4bacd043ded5e6d3c6f4dd74cdcff8da0c38fa792a8c3dbf3154040aef00e9 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.060985    4824 scope.go:117] "RemoveContainer" containerID="cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.061299    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05\": container with ID starting with cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05 not found: ID does not exist" containerID="cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.061332    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05"} err="failed to get container status \"cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05\": rpc error: code = NotFound desc = could not find container \"cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05\": container with ID starting with cb46f852b1fa95f55ef928affdef17db3a7757f10b86fdac1b18139489ca9a05 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.061349    4824 scope.go:117] "RemoveContainer" containerID="02303037a8aadc54e488e0b6a7ab3aa3d1e96d811720314cf37ff8918ad099fb"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.081911    4824 scope.go:117] "RemoveContainer" containerID="6e4723e09d2c90df1133040b661458e056a5c4c1f00d93bf75e8c99376670084"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.107205    4824 scope.go:117] "RemoveContainer" containerID="d2bddba741f540f51d957b8348722b2dd5b6390ad7a31df14141f87d367c925f"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.125468    4824 scope.go:117] "RemoveContainer" containerID="ec8e75cf7df9b51326581fdd0a2eaea619efb86432955ae4d521989852d32634"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.142077    4824 scope.go:117] "RemoveContainer" containerID="69fc867d077f92fdd2db80cceb03a2e106a1ea50be845d7d9442fb248de05920"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.160334    4824 scope.go:117] "RemoveContainer" containerID="283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.176764    4824 scope.go:117] "RemoveContainer" containerID="e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.199686    4824 scope.go:117] "RemoveContainer" containerID="d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.228807    4824 scope.go:117] "RemoveContainer" containerID="0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.240273    4824 scope.go:117] "RemoveContainer" containerID="49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.240614    4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.242425    4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.242810    4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.243244    4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.243780    4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.243829    4824 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.244504    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="200ms"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.254971    4824 scope.go:117] "RemoveContainer" containerID="db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.276760    4824 scope.go:117] "RemoveContainer" containerID="283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.277139    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\": container with ID starting with 283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12 not found: ID does not exist" containerID="283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.277167    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12"} err="failed to get container status \"283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\": rpc error: code = NotFound desc = could not find container \"283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12\": container with ID starting with 283373934a6734669e193a8166fecd757b0fa2bf2e9e7c6b5caa44a8be759c12 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.277189    4824 scope.go:117] "RemoveContainer" containerID="e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.277483    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\": container with ID starting with e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335 not found: ID does not exist" containerID="e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.277504    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335"} err="failed to get container status \"e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\": rpc error: code = NotFound desc = could not find container \"e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335\": container with ID starting with e86a3843146bfd297b9fad6e2ca5baff2cdeb89bd4a3acf30f89fb56968e4335 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.277518    4824 scope.go:117] "RemoveContainer" containerID="d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.277859    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\": container with ID starting with d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4 not found: ID does not exist" containerID="d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.277994    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4"} err="failed to get container status \"d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\": rpc error: code = NotFound desc = could not find container \"d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4\": container with ID starting with d351c5e3e1567ee7ed3d52ce6a1abe8a953533bf08e0d786779baaf0f38e40f4 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.278105    4824 scope.go:117] "RemoveContainer" containerID="0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.278660    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\": container with ID starting with 0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f not found: ID does not exist" containerID="0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.278774    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f"} err="failed to get container status \"0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\": rpc error: code = NotFound desc = could not find container \"0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f\": container with ID starting with 0805121a3c7c8b8654b6719c3a6d3b1a77b393e7f9b94ef82231d172df69a85f not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.278863    4824 scope.go:117] "RemoveContainer" containerID="49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.279230    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\": container with ID starting with 49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970 not found: ID does not exist" containerID="49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.279607    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970"} err="failed to get container status \"49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\": rpc error: code = NotFound desc = could not find container \"49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970\": container with ID starting with 49b335a35751441620fc030896dccaea38a8287a73dd623038daf9b2a212e970 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.279707    4824 scope.go:117] "RemoveContainer" containerID="db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.280073    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\": container with ID starting with db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069 not found: ID does not exist" containerID="db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.280095    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069"} err="failed to get container status \"db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\": rpc error: code = NotFound desc = could not find container \"db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069\": container with ID starting with db69e5250321c143bed1e9f7d9aac6bfc8bfb94e9e3ab6e9c4b1972d6729f069 not found: ID does not exist"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.406231    4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.406642    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.406996    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.407321    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.407688    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.408026    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.408330    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:07 crc kubenswrapper[4824]: I0122 09:47:07.412106    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.445845    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="400ms"
Jan 22 09:47:07 crc kubenswrapper[4824]: E0122 09:47:07.846433    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="800ms"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.224217    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.225171    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.225481    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.225724    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.226019    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.226301    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.226580    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.267936    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kubelet-dir\") pod \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") "
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.267994    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-var-lock\") pod \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") "
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.268026    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kube-api-access\") pod \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\" (UID: \"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277\") "
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.268158    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" (UID: "1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.268185    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-var-lock" (OuterVolumeSpecName: "var-lock") pod "1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" (UID: "1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.268546    4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kubelet-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.268570    4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-var-lock\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.272962    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" (UID: "1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:47:08 crc kubenswrapper[4824]: I0122 09:47:08.370088    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277-kube-api-access\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:08 crc kubenswrapper[4824]: E0122 09:47:08.649152    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="1.6s"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.006192    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277","Type":"ContainerDied","Data":"4acf91c7aad6ece55a8e1ac7ff7d965a0cef6a38db866b15a05da7faca6fe939"}
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.006249    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4acf91c7aad6ece55a8e1ac7ff7d965a0cef6a38db866b15a05da7faca6fe939"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.006329    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.034212    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.034794    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.035328    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.035672    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.035987    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:09 crc kubenswrapper[4824]: I0122 09:47:09.036314    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:10 crc kubenswrapper[4824]: E0122 09:47:10.250901    4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="3.2s"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.405193    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.406440    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.406959    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.407435    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.411697    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.412090    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.414506    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.430027    4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.430080    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:12 crc kubenswrapper[4824]: E0122 09:47:12.430670    4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:12 crc kubenswrapper[4824]: I0122 09:47:12.432260    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:12 crc kubenswrapper[4824]: W0122 09:47:12.466249    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-58d37e193c3411f2355cd2c3e11c00f2d4ab6a2dee8b80137b658a470b022c1e WatchSource:0}: Error finding container 58d37e193c3411f2355cd2c3e11c00f2d4ab6a2dee8b80137b658a470b022c1e: Status 404 returned error can't find the container with id 58d37e193c3411f2355cd2c3e11c00f2d4ab6a2dee8b80137b658a470b022c1e
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.027367    4824 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b7d88ca4703b22ba95913ef1ba6396298c26d32b09acd67041ab680a306ac3c5" exitCode=0
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.027477    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b7d88ca4703b22ba95913ef1ba6396298c26d32b09acd67041ab680a306ac3c5"}
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.027623    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"58d37e193c3411f2355cd2c3e11c00f2d4ab6a2dee8b80137b658a470b022c1e"}
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.027869    4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.027882    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.028333    4824 status_manager.go:851] "Failed to get status for pod" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" pod="openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-77f58dc8c-h4mhx\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:13 crc kubenswrapper[4824]: E0122 09:47:13.028380    4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.028733    4824 status_manager.go:851] "Failed to get status for pod" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.029290    4824 status_manager.go:851] "Failed to get status for pod" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" pod="openshift-marketplace/redhat-marketplace-hv6zw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-hv6zw\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.029731    4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.029987    4824 status_manager.go:851] "Failed to get status for pod" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" pod="openshift-controller-manager/controller-manager-9744fd7b7-svt7c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-9744fd7b7-svt7c\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:13 crc kubenswrapper[4824]: I0122 09:47:13.030380    4824 status_manager.go:851] "Failed to get status for pod" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" pod="openshift-marketplace/community-operators-9wlq2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9wlq2\": dial tcp 38.102.83.251:6443: connect: connection refused"
Jan 22 09:47:14 crc kubenswrapper[4824]: I0122 09:47:14.034867    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6545a1211ec26a3b6f8de512408cbcb5cc9feee5e426a8bd4f67ba3c446152fc"}
Jan 22 09:47:14 crc kubenswrapper[4824]: I0122 09:47:14.035303    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c8e417ec94ffbec28231a951893f34274fa286b1cce6a7fb80abed6a70d55445"}
Jan 22 09:47:14 crc kubenswrapper[4824]: I0122 09:47:14.035313    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0aef4b7be303a55fe33ca3b6c9c8ba0ee8ff416dc471412f57fd2692238daf35"}
Jan 22 09:47:14 crc kubenswrapper[4824]: I0122 09:47:14.035322    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"81107601c49379abeeb4cd02833b51df28a1e53bbe4b18c66cea2314f956cb46"}
Jan 22 09:47:15 crc kubenswrapper[4824]: I0122 09:47:15.043170    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a6b370836f39ff832fa9a64541593cb560c68661c6c37c8c0603184116cefd5b"}
Jan 22 09:47:15 crc kubenswrapper[4824]: I0122 09:47:15.043753    4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:15 crc kubenswrapper[4824]: I0122 09:47:15.043885    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:15 crc kubenswrapper[4824]: I0122 09:47:15.044071    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:15 crc kubenswrapper[4824]: I0122 09:47:15.500219    4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body=
Jan 22 09:47:15 crc kubenswrapper[4824]: I0122 09:47:15.500280    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused"
Jan 22 09:47:16 crc kubenswrapper[4824]: I0122 09:47:16.069596    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log"
Jan 22 09:47:16 crc kubenswrapper[4824]: I0122 09:47:16.069645    4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff" exitCode=1
Jan 22 09:47:16 crc kubenswrapper[4824]: I0122 09:47:16.069677    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff"}
Jan 22 09:47:16 crc kubenswrapper[4824]: I0122 09:47:16.070118    4824 scope.go:117] "RemoveContainer" containerID="b1ce393594ff5f3f79558ee950d257b46cfc8e2845f7e3b153f89498617ec7ff"
Jan 22 09:47:17 crc kubenswrapper[4824]: I0122 09:47:17.080937    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log"
Jan 22 09:47:17 crc kubenswrapper[4824]: I0122 09:47:17.081273    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"455432b75a702abf558d8ea0d91b7dd7443529598ce57b7e83b6a9d30be3fa3c"}
Jan 22 09:47:17 crc kubenswrapper[4824]: I0122 09:47:17.433294    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:17 crc kubenswrapper[4824]: I0122 09:47:17.433668    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:17 crc kubenswrapper[4824]: I0122 09:47:17.438997    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:20 crc kubenswrapper[4824]: I0122 09:47:20.052444    4824 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:20 crc kubenswrapper[4824]: I0122 09:47:20.126682    4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:20 crc kubenswrapper[4824]: I0122 09:47:20.126714    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:20 crc kubenswrapper[4824]: I0122 09:47:20.130494    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:20 crc kubenswrapper[4824]: I0122 09:47:20.232670    4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="1cdff433-5616-406c-b493-a0f64d46377c"
Jan 22 09:47:21 crc kubenswrapper[4824]: I0122 09:47:21.136015    4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:21 crc kubenswrapper[4824]: I0122 09:47:21.136386    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:21 crc kubenswrapper[4824]: I0122 09:47:21.140196    4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="1cdff433-5616-406c-b493-a0f64d46377c"
Jan 22 09:47:25 crc kubenswrapper[4824]: I0122 09:47:25.500508    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:47:25 crc kubenswrapper[4824]: I0122 09:47:25.780304    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:47:25 crc kubenswrapper[4824]: I0122 09:47:25.785949    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.168739    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.215266    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.284879    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.557159    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.650035    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.672104    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt"
Jan 22 09:47:26 crc kubenswrapper[4824]: I0122 09:47:26.934329    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt"
Jan 22 09:47:27 crc kubenswrapper[4824]: I0122 09:47:27.032621    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt"
Jan 22 09:47:27 crc kubenswrapper[4824]: I0122 09:47:27.726596    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt"
Jan 22 09:47:27 crc kubenswrapper[4824]: I0122 09:47:27.901552    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.067514    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.174081    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.180619    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.237207    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.334045    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.393165    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt"
Jan 22 09:47:28 crc kubenswrapper[4824]: I0122 09:47:28.660686    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1"
Jan 22 09:47:29 crc kubenswrapper[4824]: I0122 09:47:29.134124    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert"
Jan 22 09:47:29 crc kubenswrapper[4824]: I0122 09:47:29.183335    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy"
Jan 22 09:47:29 crc kubenswrapper[4824]: I0122 09:47:29.708027    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config"
Jan 22 09:47:30 crc kubenswrapper[4824]: I0122 09:47:30.505133    4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160
Jan 22 09:47:30 crc kubenswrapper[4824]: I0122 09:47:30.922491    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.070742    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.358318    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.373880    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.648336    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.679419    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.691026    4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.772642    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq"
Jan 22 09:47:31 crc kubenswrapper[4824]: I0122 09:47:31.908197    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.312272    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.376662    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.613691    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.714510    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.770550    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.775909    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert"
Jan 22 09:47:32 crc kubenswrapper[4824]: I0122 09:47:32.846510    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt"
Jan 22 09:47:33 crc kubenswrapper[4824]: I0122 09:47:33.286024    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token"
Jan 22 09:47:33 crc kubenswrapper[4824]: I0122 09:47:33.286896    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config"
Jan 22 09:47:33 crc kubenswrapper[4824]: I0122 09:47:33.488134    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt"
Jan 22 09:47:33 crc kubenswrapper[4824]: I0122 09:47:33.603997    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw"
Jan 22 09:47:33 crc kubenswrapper[4824]: I0122 09:47:33.811084    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle"
Jan 22 09:47:33 crc kubenswrapper[4824]: I0122 09:47:33.873772    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.072709    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.081674    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.106044    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.189033    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.411912    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.445704    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.490673    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.751565    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.787732    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt"
Jan 22 09:47:34 crc kubenswrapper[4824]: I0122 09:47:34.947738    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.078194    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.114239    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.175597    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.199894    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.207492    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.374351    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.406956    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.580475    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.836840    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.874969    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt"
Jan 22 09:47:35 crc kubenswrapper[4824]: I0122 09:47:35.974399    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.039115    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.058337    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.140534    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.160841    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.166024    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.320003    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.326720    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.362410    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.421709    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.423253    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.434818    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.459203    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.599282    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.707550    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.772897    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.823162    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.859840    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection"
Jan 22 09:47:36 crc kubenswrapper[4824]: I0122 09:47:36.987665    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.052301    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.112289    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.172315    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.236061    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.293908    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.298953    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.311387    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.490698    4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.519632    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.542555    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.565426    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.601315    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.635928    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.792602    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.792765    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.875549    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.927860    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.935317    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff"
Jan 22 09:47:37 crc kubenswrapper[4824]: I0122 09:47:37.947921    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.019796    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.105395    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.152733    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.327376    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.391612    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.423405    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.456605    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.549028    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.563432    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.594093    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.598407    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.600479    4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.601221    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.601206664 podStartE2EDuration="38.601206664s" podCreationTimestamp="2026-01-22 09:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:47:20.088662305 +0000 UTC m=+272.992139298" watchObservedRunningTime="2026-01-22 09:47:38.601206664 +0000 UTC m=+291.504683657"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604307    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wlq2","openshift-route-controller-manager/route-controller-manager-77f58dc8c-h4mhx","openshift-controller-manager/controller-manager-9744fd7b7-svt7c","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-marketplace-hv6zw"]
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604390    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-788989cbfd-d5xhw","openshift-kube-apiserver/kube-apiserver-crc","openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"]
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604779    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" containerName="controller-manager"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604799    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" containerName="controller-manager"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604832    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="registry-server"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604842    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="registry-server"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604858    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="registry-server"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604887    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="registry-server"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604899    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="extract-utilities"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604908    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="extract-utilities"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604918    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="extract-content"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604947    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="extract-content"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604963    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" containerName="installer"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604971    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" containerName="installer"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604983    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="extract-content"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.604990    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="extract-content"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.604999    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="extract-utilities"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605005    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="extract-utilities"
Jan 22 09:47:38 crc kubenswrapper[4824]: E0122 09:47:38.605015    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" containerName="route-controller-manager"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605021    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" containerName="route-controller-manager"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605124    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" containerName="registry-server"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605135    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e4e9337-a9f7-42b2-b7a4-5cd46ae9b277" containerName="installer"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605143    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" containerName="route-controller-manager"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605156    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" containerName="registry-server"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605164    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" containerName="controller-manager"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605713    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605926    4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.605968    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d9e5d7f0-3820-4ff5-8141-e66990987acf"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.606266    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-822xl"]
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.606322    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.612124    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.612608    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.612789    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.612921    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613051    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613077    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613161    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613168    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613218    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613250    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613430    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.613818    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.614222    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.616545    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.639611    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.639874    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.668537    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.686989    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.686969657 podStartE2EDuration="18.686969657s" podCreationTimestamp="2026-01-22 09:47:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:47:38.683007255 +0000 UTC m=+291.586484248" watchObservedRunningTime="2026-01-22 09:47:38.686969657 +0000 UTC m=+291.590446650"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.718135    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757619    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-client-ca\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757683    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-proxy-ca-bundles\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757708    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846aee0b-b33c-433f-bdec-4eca96ca16ad-serving-cert\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757724    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c169da9-4705-47be-8206-e57d64ac96ca-serving-cert\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757740    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-config\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757763    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbgpz\" (UniqueName: \"kubernetes.io/projected/846aee0b-b33c-433f-bdec-4eca96ca16ad-kube-api-access-mbgpz\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757900    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c169da9-4705-47be-8206-e57d64ac96ca-client-ca\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.757966    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg2fd\" (UniqueName: \"kubernetes.io/projected/2c169da9-4705-47be-8206-e57d64ac96ca-kube-api-access-dg2fd\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.758114    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c169da9-4705-47be-8206-e57d64ac96ca-config\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.806405    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.853348    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.859106    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-client-ca\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.859427    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-proxy-ca-bundles\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.859554    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846aee0b-b33c-433f-bdec-4eca96ca16ad-serving-cert\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.859649    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c169da9-4705-47be-8206-e57d64ac96ca-serving-cert\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.859757    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-config\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.859973    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbgpz\" (UniqueName: \"kubernetes.io/projected/846aee0b-b33c-433f-bdec-4eca96ca16ad-kube-api-access-mbgpz\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.860086    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c169da9-4705-47be-8206-e57d64ac96ca-client-ca\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.860182    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-client-ca\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.860200    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg2fd\" (UniqueName: \"kubernetes.io/projected/2c169da9-4705-47be-8206-e57d64ac96ca-kube-api-access-dg2fd\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.860453    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c169da9-4705-47be-8206-e57d64ac96ca-config\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.860917    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-proxy-ca-bundles\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.861568    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/846aee0b-b33c-433f-bdec-4eca96ca16ad-config\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.861584    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c169da9-4705-47be-8206-e57d64ac96ca-client-ca\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.862010    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c169da9-4705-47be-8206-e57d64ac96ca-config\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.871976    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c169da9-4705-47be-8206-e57d64ac96ca-serving-cert\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.873246    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/846aee0b-b33c-433f-bdec-4eca96ca16ad-serving-cert\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.880003    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg2fd\" (UniqueName: \"kubernetes.io/projected/2c169da9-4705-47be-8206-e57d64ac96ca-kube-api-access-dg2fd\") pod \"route-controller-manager-7948b96f46-9qfx2\" (UID: \"2c169da9-4705-47be-8206-e57d64ac96ca\") " pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.882586    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbgpz\" (UniqueName: \"kubernetes.io/projected/846aee0b-b33c-433f-bdec-4eca96ca16ad-kube-api-access-mbgpz\") pod \"controller-manager-788989cbfd-d5xhw\" (UID: \"846aee0b-b33c-433f-bdec-4eca96ca16ad\") " pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.903215    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.944278    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:38 crc kubenswrapper[4824]: I0122 09:47:38.952733    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.005238    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.081747    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.103171    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.112434    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.130464    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.208240    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.209144    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.230808    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.386844    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.411583    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd47eee-b6b0-4448-8bcd-09c188762714" path="/var/lib/kubelet/pods/2fd47eee-b6b0-4448-8bcd-09c188762714/volumes"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.412603    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3860e075-2372-4e81-aece-e4bb5ff66e7e" path="/var/lib/kubelet/pods/3860e075-2372-4e81-aece-e4bb5ff66e7e/volumes"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.413114    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce653fe-67cf-4a94-bff2-cf4ef16b4e38" path="/var/lib/kubelet/pods/4ce653fe-67cf-4a94-bff2-cf4ef16b4e38/volumes"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.414103    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe569b9-ef9e-4ce9-b985-511797cc0633" path="/var/lib/kubelet/pods/cfe569b9-ef9e-4ce9-b985-511797cc0633/volumes"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.536402    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.577094    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.614951    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.624429    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.759387    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.815420    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.848979    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.877804    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.956871    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.971888    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-788989cbfd-d5xhw"]
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.981719    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.987092    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt"
Jan 22 09:47:39 crc kubenswrapper[4824]: I0122 09:47:39.990443    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"]
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.170047    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.330182    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.353740    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.536977    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.642637    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.695625    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.716472    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.725908    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.782100    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.819485    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.864028    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.872255    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls"
Jan 22 09:47:40 crc kubenswrapper[4824]: I0122 09:47:40.965336    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.009257    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.041467    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.082633    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.119206    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.159315    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.271678    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.362029    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.400670    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.439514    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.448483    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.458725    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.560332    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca"
Jan 22 09:47:41 crc kubenswrapper[4824]: I0122 09:47:41.621577    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt"
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.096408    4824 log.go:32] "RunPodSandbox from runtime service failed" err=<
Jan 22 09:47:42 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e" Netns:"/var/run/netns/973a9fab-0318-41e6-9e60-2f2c090f69ff" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod "route-controller-manager-7948b96f46-9qfx2" not found
Jan 22 09:47:42 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:42 crc kubenswrapper[4824]:  >
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.096485    4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=<
Jan 22 09:47:42 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e" Netns:"/var/run/netns/973a9fab-0318-41e6-9e60-2f2c090f69ff" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod "route-controller-manager-7948b96f46-9qfx2" not found
Jan 22 09:47:42 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:42 crc kubenswrapper[4824]:  > pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.096506    4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=<
Jan 22 09:47:42 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e" Netns:"/var/run/netns/973a9fab-0318-41e6-9e60-2f2c090f69ff" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod "route-controller-manager-7948b96f46-9qfx2" not found
Jan 22 09:47:42 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:42 crc kubenswrapper[4824]:  > pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.096567    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager(2c169da9-4705-47be-8206-e57d64ac96ca)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager(2c169da9-4705-47be-8206-e57d64ac96ca)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e\\\" Netns:\\\"/var/run/netns/973a9fab-0318-41e6-9e60-2f2c090f69ff\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=1617b21f62eeb328c57b595410a9cab2f8b7f8156cd80f07702f772ec9b9d47e;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod \\\"route-controller-manager-7948b96f46-9qfx2\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2" podUID="2c169da9-4705-47be-8206-e57d64ac96ca"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.121748    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config"
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.143844    4824 log.go:32] "RunPodSandbox from runtime service failed" err=<
Jan 22 09:47:42 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f" Netns:"/var/run/netns/66ef417f-6a04-4a44-b159-7fa40e167178" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod "controller-manager-788989cbfd-d5xhw" not found
Jan 22 09:47:42 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:42 crc kubenswrapper[4824]:  >
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.143918    4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=<
Jan 22 09:47:42 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f" Netns:"/var/run/netns/66ef417f-6a04-4a44-b159-7fa40e167178" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod "controller-manager-788989cbfd-d5xhw" not found
Jan 22 09:47:42 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:42 crc kubenswrapper[4824]:  > pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.143943    4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=<
Jan 22 09:47:42 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f" Netns:"/var/run/netns/66ef417f-6a04-4a44-b159-7fa40e167178" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod "controller-manager-788989cbfd-d5xhw" not found
Jan 22 09:47:42 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:42 crc kubenswrapper[4824]:  > pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:42 crc kubenswrapper[4824]: E0122 09:47:42.144015    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-788989cbfd-d5xhw_openshift-controller-manager(846aee0b-b33c-433f-bdec-4eca96ca16ad)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-788989cbfd-d5xhw_openshift-controller-manager(846aee0b-b33c-433f-bdec-4eca96ca16ad)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f\\\" Netns:\\\"/var/run/netns/66ef417f-6a04-4a44-b159-7fa40e167178\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=8163da937842dd9c897877871d1993cccfc8c0bd4c72c34a2cf0168ea756431f;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod \\\"controller-manager-788989cbfd-d5xhw\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw" podUID="846aee0b-b33c-433f-bdec-4eca96ca16ad"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.253441    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.271600    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.271656    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.272049    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.272164    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.299558    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.378065    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.406316    4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"]
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.406631    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425" gracePeriod=5
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.410344    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.411587    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.442841    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.505463    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.657977    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin"
Jan 22 09:47:42 crc kubenswrapper[4824]: I0122 09:47:42.669036    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.040703    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.077240    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.140652    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.239100    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.259876    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.447091    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.457297    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.492315    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.688826    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.733066    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.741286    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.750036    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.787520    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.895293    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.918820    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.984223    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.985088    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert"
Jan 22 09:47:43 crc kubenswrapper[4824]: I0122 09:47:43.986370    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.140985    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.181174    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.269441    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.286418    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.316387    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.341515    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.375977    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.414857    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.555768    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.670012    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.896724    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.955812    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images"
Jan 22 09:47:44 crc kubenswrapper[4824]: I0122 09:47:44.971435    4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.059982    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert"
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.203456    4824 log.go:32] "RunPodSandbox from runtime service failed" err=<
Jan 22 09:47:45 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025" Netns:"/var/run/netns/aaac6a40-1825-4f0c-8d03-d01e16069c3f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod "route-controller-manager-7948b96f46-9qfx2" not found
Jan 22 09:47:45 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:45 crc kubenswrapper[4824]:  >
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.203742    4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=<
Jan 22 09:47:45 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025" Netns:"/var/run/netns/aaac6a40-1825-4f0c-8d03-d01e16069c3f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod "route-controller-manager-7948b96f46-9qfx2" not found
Jan 22 09:47:45 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:45 crc kubenswrapper[4824]:  > pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.203765    4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=<
Jan 22 09:47:45 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025" Netns:"/var/run/netns/aaac6a40-1825-4f0c-8d03-d01e16069c3f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod "route-controller-manager-7948b96f46-9qfx2" not found
Jan 22 09:47:45 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:45 crc kubenswrapper[4824]:  > pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.203830    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager(2c169da9-4705-47be-8206-e57d64ac96ca)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager(2c169da9-4705-47be-8206-e57d64ac96ca)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-7948b96f46-9qfx2_openshift-route-controller-manager_2c169da9-4705-47be-8206-e57d64ac96ca_0(58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025): error adding pod openshift-route-controller-manager_route-controller-manager-7948b96f46-9qfx2 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025\\\" Netns:\\\"/var/run/netns/aaac6a40-1825-4f0c-8d03-d01e16069c3f\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-7948b96f46-9qfx2;K8S_POD_INFRA_CONTAINER_ID=58fa702af16ede5e6f85c35d3be538873dd086cca1ef3351442d556a4ef55025;K8S_POD_UID=2c169da9-4705-47be-8206-e57d64ac96ca\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2] networking: Multus: [openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2/2c169da9-4705-47be-8206-e57d64ac96ca]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-7948b96f46-9qfx2 in out of cluster comm: pod \\\"route-controller-manager-7948b96f46-9qfx2\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2" podUID="2c169da9-4705-47be-8206-e57d64ac96ca"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.204340    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk"
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.265602    4824 log.go:32] "RunPodSandbox from runtime service failed" err=<
Jan 22 09:47:45 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687" Netns:"/var/run/netns/64102435-b95e-4ac5-8c2f-b3e8197ef0a8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod "controller-manager-788989cbfd-d5xhw" not found
Jan 22 09:47:45 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:45 crc kubenswrapper[4824]:  >
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.265697    4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=<
Jan 22 09:47:45 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687" Netns:"/var/run/netns/64102435-b95e-4ac5-8c2f-b3e8197ef0a8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod "controller-manager-788989cbfd-d5xhw" not found
Jan 22 09:47:45 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:45 crc kubenswrapper[4824]:  > pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.265732    4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=<
Jan 22 09:47:45 crc kubenswrapper[4824]:         rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687" Netns:"/var/run/netns/64102435-b95e-4ac5-8c2f-b3e8197ef0a8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod "controller-manager-788989cbfd-d5xhw" not found
Jan 22 09:47:45 crc kubenswrapper[4824]:         ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"}
Jan 22 09:47:45 crc kubenswrapper[4824]:  > pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:47:45 crc kubenswrapper[4824]: E0122 09:47:45.265819    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-788989cbfd-d5xhw_openshift-controller-manager(846aee0b-b33c-433f-bdec-4eca96ca16ad)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-788989cbfd-d5xhw_openshift-controller-manager(846aee0b-b33c-433f-bdec-4eca96ca16ad)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-788989cbfd-d5xhw_openshift-controller-manager_846aee0b-b33c-433f-bdec-4eca96ca16ad_0(f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687): error adding pod openshift-controller-manager_controller-manager-788989cbfd-d5xhw to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687\\\" Netns:\\\"/var/run/netns/64102435-b95e-4ac5-8c2f-b3e8197ef0a8\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-788989cbfd-d5xhw;K8S_POD_INFRA_CONTAINER_ID=f541f5459478c709dbef4a9ce86061d346239b7ae16984fefbc0b024da44c687;K8S_POD_UID=846aee0b-b33c-433f-bdec-4eca96ca16ad\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-788989cbfd-d5xhw] networking: Multus: [openshift-controller-manager/controller-manager-788989cbfd-d5xhw/846aee0b-b33c-433f-bdec-4eca96ca16ad]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-788989cbfd-d5xhw in out of cluster comm: pod \\\"controller-manager-788989cbfd-d5xhw\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw" podUID="846aee0b-b33c-433f-bdec-4eca96ca16ad"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.306980    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.457268    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.543202    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.653435    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.691433    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.713479    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.814203    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.910148    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default"
Jan 22 09:47:45 crc kubenswrapper[4824]: I0122 09:47:45.993550    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.015378    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.106756    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.107339    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.219273    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.442571    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.768330    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw"
Jan 22 09:47:46 crc kubenswrapper[4824]: I0122 09:47:46.770145    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert"
Jan 22 09:47:47 crc kubenswrapper[4824]: I0122 09:47:47.069808    4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k"
Jan 22 09:47:47 crc kubenswrapper[4824]: I0122 09:47:47.134311    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit"
Jan 22 09:47:47 crc kubenswrapper[4824]: I0122 09:47:47.261031    4824 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials
Jan 22 09:47:47 crc kubenswrapper[4824]: I0122 09:47:47.738582    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig"
Jan 22 09:47:47 crc kubenswrapper[4824]: I0122 09:47:47.817584    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.014467    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.014600    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.053732    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.134070    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190199    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") "
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190324    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") "
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190333    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190413    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") "
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190450    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") "
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190476    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") "
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190498    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190578    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190612    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190884    4824 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190904    4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190913    4824 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.190924    4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.200406    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.292485    4824 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.304965    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.305039    4824 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425" exitCode=137
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.305090    4824 scope.go:117] "RemoveContainer" containerID="7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.305117    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.323677    4824 scope.go:117] "RemoveContainer" containerID="7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425"
Jan 22 09:47:48 crc kubenswrapper[4824]: E0122 09:47:48.324145    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425\": container with ID starting with 7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425 not found: ID does not exist" containerID="7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.324173    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425"} err="failed to get container status \"7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425\": rpc error: code = NotFound desc = could not find container \"7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425\": container with ID starting with 7fe7f0832367a110ba2a57b646708a7f03eed999d813a66971e3492ca9b30425 not found: ID does not exist"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.351021    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.464278    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.523567    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.799683    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls"
Jan 22 09:47:48 crc kubenswrapper[4824]: I0122 09:47:48.799764    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw"
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.008388    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt"
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.129002    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd"
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.337047    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm"
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.416104    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes"
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.416694    4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID=""
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.430844    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"]
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.430887    4824 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0a0384b8-5cbd-4dc5-9347-67fa7afe5442"
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.434202    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"]
Jan 22 09:47:49 crc kubenswrapper[4824]: I0122 09:47:49.434250    4824 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0a0384b8-5cbd-4dc5-9347-67fa7afe5442"
Jan 22 09:47:50 crc kubenswrapper[4824]: I0122 09:47:50.309733    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt"
Jan 22 09:47:58 crc kubenswrapper[4824]: I0122 09:47:58.404074    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:58 crc kubenswrapper[4824]: I0122 09:47:58.404955    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:47:58 crc kubenswrapper[4824]: I0122 09:47:58.624194    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"]
Jan 22 09:47:59 crc kubenswrapper[4824]: I0122 09:47:59.379622    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2" event={"ID":"2c169da9-4705-47be-8206-e57d64ac96ca","Type":"ContainerStarted","Data":"e4384716b999e681b8839290d906b18f70c590a8fbd3c4f516215a7385fcc026"}
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.398172    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2" event={"ID":"2c169da9-4705-47be-8206-e57d64ac96ca","Type":"ContainerStarted","Data":"e5e52e8f958105f613294db7ec1033f0be22680d8e08c8c27cfa470102a0f674"}
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.398929    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.404205    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.405053    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.410546    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2"
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.454621    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7948b96f46-9qfx2" podStartSLOduration=60.454603097 podStartE2EDuration="1m0.454603097s" podCreationTimestamp="2026-01-22 09:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:00.43069472 +0000 UTC m=+313.334171713" watchObservedRunningTime="2026-01-22 09:48:00.454603097 +0000 UTC m=+313.358080080"
Jan 22 09:48:00 crc kubenswrapper[4824]: I0122 09:48:00.609898    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-788989cbfd-d5xhw"]
Jan 22 09:48:00 crc kubenswrapper[4824]: W0122 09:48:00.617314    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod846aee0b_b33c_433f_bdec_4eca96ca16ad.slice/crio-35d5f694237df1a78b8f07e7b9d3347b7414c2baf2b07471e17052132658a924 WatchSource:0}: Error finding container 35d5f694237df1a78b8f07e7b9d3347b7414c2baf2b07471e17052132658a924: Status 404 returned error can't find the container with id 35d5f694237df1a78b8f07e7b9d3347b7414c2baf2b07471e17052132658a924
Jan 22 09:48:01 crc kubenswrapper[4824]: I0122 09:48:01.410921    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:48:01 crc kubenswrapper[4824]: I0122 09:48:01.411284    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw" event={"ID":"846aee0b-b33c-433f-bdec-4eca96ca16ad","Type":"ContainerStarted","Data":"a75fb8ebb27ed8f4c5ff068cd94b08b8dff8640a4760000d3fe407a368afb645"}
Jan 22 09:48:01 crc kubenswrapper[4824]: I0122 09:48:01.411310    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw" event={"ID":"846aee0b-b33c-433f-bdec-4eca96ca16ad","Type":"ContainerStarted","Data":"35d5f694237df1a78b8f07e7b9d3347b7414c2baf2b07471e17052132658a924"}
Jan 22 09:48:01 crc kubenswrapper[4824]: I0122 09:48:01.422154    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw"
Jan 22 09:48:01 crc kubenswrapper[4824]: I0122 09:48:01.428587    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-788989cbfd-d5xhw" podStartSLOduration=61.428570954 podStartE2EDuration="1m1.428570954s" podCreationTimestamp="2026-01-22 09:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:01.426646585 +0000 UTC m=+314.330123598" watchObservedRunningTime="2026-01-22 09:48:01.428570954 +0000 UTC m=+314.332047947"
Jan 22 09:48:03 crc kubenswrapper[4824]: I0122 09:48:03.628961    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" containerName="oauth-openshift" containerID="cri-o://6e0a8d776d28a55a8220214416d4cf614c7325924f7eb84819824c6ebbcb6a02" gracePeriod=15
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.427882    4824 generic.go:334] "Generic (PLEG): container finished" podID="fc660b59-eb41-4c18-84a4-ebd782752570" containerID="6e0a8d776d28a55a8220214416d4cf614c7325924f7eb84819824c6ebbcb6a02" exitCode=0
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.427987    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" event={"ID":"fc660b59-eb41-4c18-84a4-ebd782752570","Type":"ContainerDied","Data":"6e0a8d776d28a55a8220214416d4cf614c7325924f7eb84819824c6ebbcb6a02"}
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.501010    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.527093    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"]
Jan 22 09:48:05 crc kubenswrapper[4824]: E0122 09:48:05.527333    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.527347    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor"
Jan 22 09:48:05 crc kubenswrapper[4824]: E0122 09:48:05.527380    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" containerName="oauth-openshift"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.527389    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" containerName="oauth-openshift"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.527503    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.527528    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" containerName="oauth-openshift"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.527951    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.539568    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"]
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623647    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-serving-cert\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623700    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-idp-0-file-data\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623725    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fc660b59-eb41-4c18-84a4-ebd782752570-audit-dir\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623750    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-router-certs\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623782    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-error\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623811    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt9x8\" (UniqueName: \"kubernetes.io/projected/fc660b59-eb41-4c18-84a4-ebd782752570-kube-api-access-zt9x8\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623844    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-cliconfig\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623860    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-audit-policies\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623881    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-session\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623909    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-provider-selection\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623945    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-login\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623965    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-service-ca\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.623992    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-ocp-branding-template\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624008    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-trusted-ca-bundle\") pod \"fc660b59-eb41-4c18-84a4-ebd782752570\" (UID: \"fc660b59-eb41-4c18-84a4-ebd782752570\") "
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624125    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624152    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624179    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9777dee4-65c9-4c1e-bafc-324f3bc0f393-audit-dir\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624193    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624212    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-login\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624236    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624279    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-service-ca\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624301    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-error\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624321    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-router-certs\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624338    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zblwb\" (UniqueName: \"kubernetes.io/projected/9777dee4-65c9-4c1e-bafc-324f3bc0f393-kube-api-access-zblwb\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624370    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-session\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624391    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624411    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.624428    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-audit-policies\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.625910    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.625955    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.626166    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.626212    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc660b59-eb41-4c18-84a4-ebd782752570-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.627707    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.630830    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.630840    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc660b59-eb41-4c18-84a4-ebd782752570-kube-api-access-zt9x8" (OuterVolumeSpecName: "kube-api-access-zt9x8") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "kube-api-access-zt9x8". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.631792    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.632004    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.632159    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.632350    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.644073    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.645300    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.645495    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "fc660b59-eb41-4c18-84a4-ebd782752570" (UID: "fc660b59-eb41-4c18-84a4-ebd782752570"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.725768    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9777dee4-65c9-4c1e-bafc-324f3bc0f393-audit-dir\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.725923    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.725857    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9777dee4-65c9-4c1e-bafc-324f3bc0f393-audit-dir\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726008    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-login\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726655    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726725    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-service-ca\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726759    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-error\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726833    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-router-certs\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726865    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zblwb\" (UniqueName: \"kubernetes.io/projected/9777dee4-65c9-4c1e-bafc-324f3bc0f393-kube-api-access-zblwb\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726890    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-session\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726922    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726958    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.726986    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-audit-policies\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727010    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727037    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727090    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727104    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727118    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt9x8\" (UniqueName: \"kubernetes.io/projected/fc660b59-eb41-4c18-84a4-ebd782752570-kube-api-access-zt9x8\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727132    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727146    4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-audit-policies\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727160    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-session\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727284    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727469    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727505    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727521    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727534    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727548    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727563    4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/fc660b59-eb41-4c18-84a4-ebd782752570-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.727589    4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fc660b59-eb41-4c18-84a4-ebd782752570-audit-dir\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.728478    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-service-ca\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.728572    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-audit-policies\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.729335    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.730327    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.731072    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-error\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.731470    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-login\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.732246    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-router-certs\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.732794    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.733161    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.735050    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-session\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.735416    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.735787    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9777dee4-65c9-4c1e-bafc-324f3bc0f393-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.742564    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zblwb\" (UniqueName: \"kubernetes.io/projected/9777dee4-65c9-4c1e-bafc-324f3bc0f393-kube-api-access-zblwb\") pod \"oauth-openshift-7b66cbb5c5-47xk7\" (UID: \"9777dee4-65c9-4c1e-bafc-324f3bc0f393\") " pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:05 crc kubenswrapper[4824]: I0122 09:48:05.850531    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.311333    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"]
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.436712    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-822xl" event={"ID":"fc660b59-eb41-4c18-84a4-ebd782752570","Type":"ContainerDied","Data":"fd63951e89ff43f969261b165cad15530dae87833644b179376372eecf9c8399"}
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.436775    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-822xl"
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.436794    4824 scope.go:117] "RemoveContainer" containerID="6e0a8d776d28a55a8220214416d4cf614c7325924f7eb84819824c6ebbcb6a02"
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.438951    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7" event={"ID":"9777dee4-65c9-4c1e-bafc-324f3bc0f393","Type":"ContainerStarted","Data":"768d690336ef25cf4c4fe17b9480274238e12bfcfd3f98120c54323072d8a41f"}
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.486161    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-822xl"]
Jan 22 09:48:06 crc kubenswrapper[4824]: I0122 09:48:06.490805    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-822xl"]
Jan 22 09:48:07 crc kubenswrapper[4824]: I0122 09:48:07.415596    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc660b59-eb41-4c18-84a4-ebd782752570" path="/var/lib/kubelet/pods/fc660b59-eb41-4c18-84a4-ebd782752570/volumes"
Jan 22 09:48:10 crc kubenswrapper[4824]: I0122 09:48:10.471600    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7" event={"ID":"9777dee4-65c9-4c1e-bafc-324f3bc0f393","Type":"ContainerStarted","Data":"bbbb4c830b73aa1a1da7f02911dae2e9025dd070cf7bdfb72e750e1d73cbed88"}
Jan 22 09:48:10 crc kubenswrapper[4824]: I0122 09:48:10.472074    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:10 crc kubenswrapper[4824]: I0122 09:48:10.483531    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7"
Jan 22 09:48:10 crc kubenswrapper[4824]: I0122 09:48:10.507415    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7b66cbb5c5-47xk7" podStartSLOduration=37.507349631 podStartE2EDuration="37.507349631s" podCreationTimestamp="2026-01-22 09:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:10.50665621 +0000 UTC m=+323.410133303" watchObservedRunningTime="2026-01-22 09:48:10.507349631 +0000 UTC m=+323.410826674"
Jan 22 09:48:14 crc kubenswrapper[4824]: I0122 09:48:14.989521    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qvd2r"]
Jan 22 09:48:14 crc kubenswrapper[4824]: I0122 09:48:14.990641    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qvd2r" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="registry-server" containerID="cri-o://40981f624a90205622eab9f1ca631923b8deed71d17240a64a0f9f2432fa979b" gracePeriod=30
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.000418    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9rdhx"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.000643    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9rdhx" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="registry-server" containerID="cri-o://1419fdfd5269cfff5e11ab0c006f882559be5f169910c27a72809be04a43be39" gracePeriod=30
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.010821    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7qjxk"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.011006    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator" containerID="cri-o://98fd915035c3a2f6e9220db622a5677a3e317ca2d01d7ecb30a5377bfb45b1f2" gracePeriod=30
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.020389    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfqfs"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.021588    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xfqfs" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="registry-server" containerID="cri-o://2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921" gracePeriod=30
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.025615    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkknc"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.026850    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.045543    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bkmd8"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.045761    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bkmd8" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="registry-server" containerID="cri-o://a4c02e3b803ee63ea2c2298af19e018d640051c4f1cf7b2bee77291fc7b6c940" gracePeriod=30
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.050871    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/08665d12-ff37-4d48-ae5f-caa77a4661d3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.050947    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wp5t\" (UniqueName: \"kubernetes.io/projected/08665d12-ff37-4d48-ae5f-caa77a4661d3-kube-api-access-9wp5t\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.051008    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08665d12-ff37-4d48-ae5f-caa77a4661d3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.056315    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.056632    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xfnlv" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="registry-server" containerID="cri-o://a3a6c2c04f313c45a5892849d9c800f01ecc2e84447a62a45ae85fce1d194ed4" gracePeriod=30
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.099178    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-9rdhx" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="registry-server" probeResult="failure" output=""
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.114733    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkknc"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.120021    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-9rdhx" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="registry-server" probeResult="failure" output=""
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.152005    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/08665d12-ff37-4d48-ae5f-caa77a4661d3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.152087    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wp5t\" (UniqueName: \"kubernetes.io/projected/08665d12-ff37-4d48-ae5f-caa77a4661d3-kube-api-access-9wp5t\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.152150    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08665d12-ff37-4d48-ae5f-caa77a4661d3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.153383    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/08665d12-ff37-4d48-ae5f-caa77a4661d3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.159240    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/08665d12-ff37-4d48-ae5f-caa77a4661d3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.172195    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wp5t\" (UniqueName: \"kubernetes.io/projected/08665d12-ff37-4d48-ae5f-caa77a4661d3-kube-api-access-9wp5t\") pod \"marketplace-operator-79b997595-wkknc\" (UID: \"08665d12-ff37-4d48-ae5f-caa77a4661d3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.351816    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.455735    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:48:15 crc kubenswrapper[4824]: I0122 09:48:15.792573    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkknc"]
Jan 22 09:48:15 crc kubenswrapper[4824]: W0122 09:48:15.797347    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08665d12_ff37_4d48_ae5f_caa77a4661d3.slice/crio-d10867085da5faf4d78dc3dfa8866ca1963cbe9341761b7fe80888fc3ac9b464 WatchSource:0}: Error finding container d10867085da5faf4d78dc3dfa8866ca1963cbe9341761b7fe80888fc3ac9b464: Status 404 returned error can't find the container with id d10867085da5faf4d78dc3dfa8866ca1963cbe9341761b7fe80888fc3ac9b464
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.506393    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc" event={"ID":"08665d12-ff37-4d48-ae5f-caa77a4661d3","Type":"ContainerStarted","Data":"0d18d983a2afdfd2c9b4dbba7cd076283e2a5eb26eef1b6d80f54be52d9281f7"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.506711    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc" event={"ID":"08665d12-ff37-4d48-ae5f-caa77a4661d3","Type":"ContainerStarted","Data":"d10867085da5faf4d78dc3dfa8866ca1963cbe9341761b7fe80888fc3ac9b464"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.506727    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.509236    4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wkknc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" start-of-body=
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.509302    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc" podUID="08665d12-ff37-4d48-ae5f-caa77a4661d3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused"
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.513908    4824 generic.go:334] "Generic (PLEG): container finished" podID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerID="98fd915035c3a2f6e9220db622a5677a3e317ca2d01d7ecb30a5377bfb45b1f2" exitCode=0
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.513990    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" event={"ID":"2463610b-364e-4154-83ee-b6c00a6c6da5","Type":"ContainerDied","Data":"98fd915035c3a2f6e9220db622a5677a3e317ca2d01d7ecb30a5377bfb45b1f2"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.523787    4824 generic.go:334] "Generic (PLEG): container finished" podID="2b498256-7f0d-4941-998e-e670f5941cdb" containerID="a4c02e3b803ee63ea2c2298af19e018d640051c4f1cf7b2bee77291fc7b6c940" exitCode=0
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.523873    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerDied","Data":"a4c02e3b803ee63ea2c2298af19e018d640051c4f1cf7b2bee77291fc7b6c940"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.528976    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc" podStartSLOduration=1.528948974 podStartE2EDuration="1.528948974s" podCreationTimestamp="2026-01-22 09:48:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:16.524202585 +0000 UTC m=+329.427679578" watchObservedRunningTime="2026-01-22 09:48:16.528948974 +0000 UTC m=+329.432425967"
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.530994    4824 generic.go:334] "Generic (PLEG): container finished" podID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerID="a3a6c2c04f313c45a5892849d9c800f01ecc2e84447a62a45ae85fce1d194ed4" exitCode=0
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.531071    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerDied","Data":"a3a6c2c04f313c45a5892849d9c800f01ecc2e84447a62a45ae85fce1d194ed4"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.535480    4824 generic.go:334] "Generic (PLEG): container finished" podID="072ef9f4-8754-45c5-9600-accb760d786c" containerID="1419fdfd5269cfff5e11ab0c006f882559be5f169910c27a72809be04a43be39" exitCode=0
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.535590    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rdhx" event={"ID":"072ef9f4-8754-45c5-9600-accb760d786c","Type":"ContainerDied","Data":"1419fdfd5269cfff5e11ab0c006f882559be5f169910c27a72809be04a43be39"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.539045    4824 generic.go:334] "Generic (PLEG): container finished" podID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerID="40981f624a90205622eab9f1ca631923b8deed71d17240a64a0f9f2432fa979b" exitCode=0
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.539094    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qvd2r" event={"ID":"88bd6885-a443-4d1f-b06e-5dd0f3107dab","Type":"ContainerDied","Data":"40981f624a90205622eab9f1ca631923b8deed71d17240a64a0f9f2432fa979b"}
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.541762    4824 generic.go:334] "Generic (PLEG): container finished" podID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerID="2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921" exitCode=0
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.541807    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfqfs" event={"ID":"8d2678ef-e863-4181-a497-b98b7fff0cce","Type":"ContainerDied","Data":"2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921"}
Jan 22 09:48:16 crc kubenswrapper[4824]: E0122 09:48:16.662892    4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921 is running failed: container process not found" containerID="2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921" cmd=["grpc_health_probe","-addr=:50051"]
Jan 22 09:48:16 crc kubenswrapper[4824]: E0122 09:48:16.663284    4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921 is running failed: container process not found" containerID="2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921" cmd=["grpc_health_probe","-addr=:50051"]
Jan 22 09:48:16 crc kubenswrapper[4824]: E0122 09:48:16.663799    4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921 is running failed: container process not found" containerID="2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921" cmd=["grpc_health_probe","-addr=:50051"]
Jan 22 09:48:16 crc kubenswrapper[4824]: E0122 09:48:16.663834    4824 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-xfqfs" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="registry-server"
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.768957    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.873809    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-catalog-content\") pod \"2b498256-7f0d-4941-998e-e670f5941cdb\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") "
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.873884    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88zfd\" (UniqueName: \"kubernetes.io/projected/2b498256-7f0d-4941-998e-e670f5941cdb-kube-api-access-88zfd\") pod \"2b498256-7f0d-4941-998e-e670f5941cdb\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") "
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.873919    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-utilities\") pod \"2b498256-7f0d-4941-998e-e670f5941cdb\" (UID: \"2b498256-7f0d-4941-998e-e670f5941cdb\") "
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.874769    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-utilities" (OuterVolumeSpecName: "utilities") pod "2b498256-7f0d-4941-998e-e670f5941cdb" (UID: "2b498256-7f0d-4941-998e-e670f5941cdb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.875880    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.880302    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b498256-7f0d-4941-998e-e670f5941cdb-kube-api-access-88zfd" (OuterVolumeSpecName: "kube-api-access-88zfd") pod "2b498256-7f0d-4941-998e-e670f5941cdb" (UID: "2b498256-7f0d-4941-998e-e670f5941cdb"). InnerVolumeSpecName "kube-api-access-88zfd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.974783    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-catalog-content\") pod \"8d2678ef-e863-4181-a497-b98b7fff0cce\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") "
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.974941    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-utilities\") pod \"8d2678ef-e863-4181-a497-b98b7fff0cce\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") "
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.974967    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nbgs\" (UniqueName: \"kubernetes.io/projected/8d2678ef-e863-4181-a497-b98b7fff0cce-kube-api-access-9nbgs\") pod \"8d2678ef-e863-4181-a497-b98b7fff0cce\" (UID: \"8d2678ef-e863-4181-a497-b98b7fff0cce\") "
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.975221    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88zfd\" (UniqueName: \"kubernetes.io/projected/2b498256-7f0d-4941-998e-e670f5941cdb-kube-api-access-88zfd\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.975235    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.975956    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-utilities" (OuterVolumeSpecName: "utilities") pod "8d2678ef-e863-4181-a497-b98b7fff0cce" (UID: "8d2678ef-e863-4181-a497-b98b7fff0cce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.982494    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2678ef-e863-4181-a497-b98b7fff0cce-kube-api-access-9nbgs" (OuterVolumeSpecName: "kube-api-access-9nbgs") pod "8d2678ef-e863-4181-a497-b98b7fff0cce" (UID: "8d2678ef-e863-4181-a497-b98b7fff0cce"). InnerVolumeSpecName "kube-api-access-9nbgs". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.993598    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b498256-7f0d-4941-998e-e670f5941cdb" (UID: "2b498256-7f0d-4941-998e-e670f5941cdb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:16 crc kubenswrapper[4824]: I0122 09:48:16.997595    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d2678ef-e863-4181-a497-b98b7fff0cce" (UID: "8d2678ef-e863-4181-a497-b98b7fff0cce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.076140    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.076200    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2678ef-e863-4181-a497-b98b7fff0cce-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.076214    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nbgs\" (UniqueName: \"kubernetes.io/projected/8d2678ef-e863-4181-a497-b98b7fff0cce-kube-api-access-9nbgs\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.076229    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b498256-7f0d-4941-998e-e670f5941cdb-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.550568    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfqfs" event={"ID":"8d2678ef-e863-4181-a497-b98b7fff0cce","Type":"ContainerDied","Data":"68395353640e3ee79eca6e1f24ad37e6e0a60966ea8e738e92a9af65c677747b"}
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.550631    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfqfs"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.550789    4824 scope.go:117] "RemoveContainer" containerID="2b41135833983d2da9af19608e4aed88cb2d293925d51d3bf6a4e16db0c60921"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.555558    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bkmd8"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.556234    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bkmd8" event={"ID":"2b498256-7f0d-4941-998e-e670f5941cdb","Type":"ContainerDied","Data":"89aebd8e5cf102fe49d7ad6910519189445e61b94c44aac76cb2dd657f817639"}
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.559595    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wkknc"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.581406    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfqfs"]
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.581587    4824 scope.go:117] "RemoveContainer" containerID="307754dde20943c0e432815035492fac2217bcd95c9dbd5f59d30bf021e0baba"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.588856    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfqfs"]
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.603148    4824 scope.go:117] "RemoveContainer" containerID="5f4a46dd06222cb576b91f5e069fc6c73acbbfefc9b525b8a75e4ea71d2dc2ba"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.613434    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bkmd8"]
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.616507    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bkmd8"]
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.629787    4824 scope.go:117] "RemoveContainer" containerID="a4c02e3b803ee63ea2c2298af19e018d640051c4f1cf7b2bee77291fc7b6c940"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.654258    4824 scope.go:117] "RemoveContainer" containerID="b97850d09dc3112ae65bb9eddd3293cd1e260d40f07ef8d03d8f610789ca85bb"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.674512    4824 scope.go:117] "RemoveContainer" containerID="02289cdb515a3089d234977eda3d52348ec5a570a77a9632cf626c40ff2ca828"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.799531    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855604    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cbw6w"]
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855854    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="extract-content"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855868    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="extract-content"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855882    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="extract-content"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855890    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="extract-content"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855899    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="extract-content"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855907    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="extract-content"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855919    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855926    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855936    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855943    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855953    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="extract-utilities"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855959    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="extract-utilities"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855970    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="extract-utilities"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855976    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="extract-utilities"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.855992    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="extract-utilities"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.855999    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="extract-utilities"
Jan 22 09:48:17 crc kubenswrapper[4824]: E0122 09:48:17.856009    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.856015    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.856132    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.856142    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.856154    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" containerName="registry-server"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.858198    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.912257    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cbw6w"]
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.988914    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-utilities\") pod \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") "
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.988976    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-catalog-content\") pod \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") "
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989128    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kcw2\" (UniqueName: \"kubernetes.io/projected/88bd6885-a443-4d1f-b06e-5dd0f3107dab-kube-api-access-8kcw2\") pod \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\" (UID: \"88bd6885-a443-4d1f-b06e-5dd0f3107dab\") "
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989304    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-bound-sa-token\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989391    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-trusted-ca\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989410    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989449    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-registry-certificates\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989468    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kl2d\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-kube-api-access-4kl2d\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989492    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-registry-tls\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989548    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989594    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.989592    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-utilities" (OuterVolumeSpecName: "utilities") pod "88bd6885-a443-4d1f-b06e-5dd0f3107dab" (UID: "88bd6885-a443-4d1f-b06e-5dd0f3107dab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:17 crc kubenswrapper[4824]: I0122 09:48:17.990190    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.009603    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88bd6885-a443-4d1f-b06e-5dd0f3107dab-kube-api-access-8kcw2" (OuterVolumeSpecName: "kube-api-access-8kcw2") pod "88bd6885-a443-4d1f-b06e-5dd0f3107dab" (UID: "88bd6885-a443-4d1f-b06e-5dd0f3107dab"). InnerVolumeSpecName "kube-api-access-8kcw2". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.010548    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.038631    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88bd6885-a443-4d1f-b06e-5dd0f3107dab" (UID: "88bd6885-a443-4d1f-b06e-5dd0f3107dab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.062565    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092088    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-bound-sa-token\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092168    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-trusted-ca\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092188    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092226    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-registry-certificates\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092243    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kl2d\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-kube-api-access-4kl2d\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092260    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-registry-tls\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092310    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092373    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kcw2\" (UniqueName: \"kubernetes.io/projected/88bd6885-a443-4d1f-b06e-5dd0f3107dab-kube-api-access-8kcw2\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092384    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88bd6885-a443-4d1f-b06e-5dd0f3107dab-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.092892    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.094997    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-trusted-ca\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.096132    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-registry-certificates\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.101167    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-registry-tls\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.108155    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-bound-sa-token\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.111311    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kl2d\" (UniqueName: \"kubernetes.io/projected/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-kube-api-access-4kl2d\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.111944    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e2af3857-54b3-4f97-9b57-c8b2afdcaaa6-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cbw6w\" (UID: \"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6\") " pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.186849    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.194326    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgxdf\" (UniqueName: \"kubernetes.io/projected/05d38c4c-ca18-478a-bcd5-01800ded772a-kube-api-access-qgxdf\") pod \"05d38c4c-ca18-478a-bcd5-01800ded772a\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.194398    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-utilities\") pod \"05d38c4c-ca18-478a-bcd5-01800ded772a\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.194430    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-catalog-content\") pod \"05d38c4c-ca18-478a-bcd5-01800ded772a\" (UID: \"05d38c4c-ca18-478a-bcd5-01800ded772a\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.195341    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-utilities" (OuterVolumeSpecName: "utilities") pod "05d38c4c-ca18-478a-bcd5-01800ded772a" (UID: "05d38c4c-ca18-478a-bcd5-01800ded772a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.198468    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05d38c4c-ca18-478a-bcd5-01800ded772a-kube-api-access-qgxdf" (OuterVolumeSpecName: "kube-api-access-qgxdf") pod "05d38c4c-ca18-478a-bcd5-01800ded772a" (UID: "05d38c4c-ca18-478a-bcd5-01800ded772a"). InnerVolumeSpecName "kube-api-access-qgxdf". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.277081    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.290390    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.298903    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgxdf\" (UniqueName: \"kubernetes.io/projected/05d38c4c-ca18-478a-bcd5-01800ded772a-kube-api-access-qgxdf\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.298941    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.343015    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05d38c4c-ca18-478a-bcd5-01800ded772a" (UID: "05d38c4c-ca18-478a-bcd5-01800ded772a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.399845    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdkr9\" (UniqueName: \"kubernetes.io/projected/072ef9f4-8754-45c5-9600-accb760d786c-kube-api-access-xdkr9\") pod \"072ef9f4-8754-45c5-9600-accb760d786c\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.399898    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-catalog-content\") pod \"072ef9f4-8754-45c5-9600-accb760d786c\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.399926    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-operator-metrics\") pod \"2463610b-364e-4154-83ee-b6c00a6c6da5\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.399975    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-utilities\") pod \"072ef9f4-8754-45c5-9600-accb760d786c\" (UID: \"072ef9f4-8754-45c5-9600-accb760d786c\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.400679    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-trusted-ca\") pod \"2463610b-364e-4154-83ee-b6c00a6c6da5\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.400717    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwkrp\" (UniqueName: \"kubernetes.io/projected/2463610b-364e-4154-83ee-b6c00a6c6da5-kube-api-access-zwkrp\") pod \"2463610b-364e-4154-83ee-b6c00a6c6da5\" (UID: \"2463610b-364e-4154-83ee-b6c00a6c6da5\") "
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.400875    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-utilities" (OuterVolumeSpecName: "utilities") pod "072ef9f4-8754-45c5-9600-accb760d786c" (UID: "072ef9f4-8754-45c5-9600-accb760d786c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.401123    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d38c4c-ca18-478a-bcd5-01800ded772a-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.401160    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.401679    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2463610b-364e-4154-83ee-b6c00a6c6da5" (UID: "2463610b-364e-4154-83ee-b6c00a6c6da5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.403891    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2463610b-364e-4154-83ee-b6c00a6c6da5" (UID: "2463610b-364e-4154-83ee-b6c00a6c6da5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.413532    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072ef9f4-8754-45c5-9600-accb760d786c-kube-api-access-xdkr9" (OuterVolumeSpecName: "kube-api-access-xdkr9") pod "072ef9f4-8754-45c5-9600-accb760d786c" (UID: "072ef9f4-8754-45c5-9600-accb760d786c"). InnerVolumeSpecName "kube-api-access-xdkr9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.415632    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2463610b-364e-4154-83ee-b6c00a6c6da5-kube-api-access-zwkrp" (OuterVolumeSpecName: "kube-api-access-zwkrp") pod "2463610b-364e-4154-83ee-b6c00a6c6da5" (UID: "2463610b-364e-4154-83ee-b6c00a6c6da5"). InnerVolumeSpecName "kube-api-access-zwkrp". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.451410    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "072ef9f4-8754-45c5-9600-accb760d786c" (UID: "072ef9f4-8754-45c5-9600-accb760d786c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.502125    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwkrp\" (UniqueName: \"kubernetes.io/projected/2463610b-364e-4154-83ee-b6c00a6c6da5-kube-api-access-zwkrp\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.502159    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdkr9\" (UniqueName: \"kubernetes.io/projected/072ef9f4-8754-45c5-9600-accb760d786c-kube-api-access-xdkr9\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.502173    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/072ef9f4-8754-45c5-9600-accb760d786c-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.502182    4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.502191    4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2463610b-364e-4154-83ee-b6c00a6c6da5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.564105    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.564076    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" event={"ID":"2463610b-364e-4154-83ee-b6c00a6c6da5","Type":"ContainerDied","Data":"dfcdc054935a47b5bd5618fc0889b41ccbfa317645c7dad101da36f76d555002"}
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.564265    4824 scope.go:117] "RemoveContainer" containerID="98fd915035c3a2f6e9220db622a5677a3e317ca2d01d7ecb30a5377bfb45b1f2"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.569980    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xfnlv" event={"ID":"05d38c4c-ca18-478a-bcd5-01800ded772a","Type":"ContainerDied","Data":"bdc553235dfbef374e8a036f93097d785519ebdd499edd1a57480c3b17ce406a"}
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.570145    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xfnlv"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.580738    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rdhx" event={"ID":"072ef9f4-8754-45c5-9600-accb760d786c","Type":"ContainerDied","Data":"a66c02577ce5c9068736f7d9f877b8f61e9940da717b3693ce322dd07db5d237"}
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.580770    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rdhx"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.584898    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qvd2r" event={"ID":"88bd6885-a443-4d1f-b06e-5dd0f3107dab","Type":"ContainerDied","Data":"0bc12fab86f9e0a51fcf153c6a77997b661babe9666acb1fd1935a94d134989c"}
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.584987    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qvd2r"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.590835    4824 scope.go:117] "RemoveContainer" containerID="a3a6c2c04f313c45a5892849d9c800f01ecc2e84447a62a45ae85fce1d194ed4"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.615993    4824 scope.go:117] "RemoveContainer" containerID="03e91e93ac0e8925bf311e95107dd5dcad5fe1e95f212a41ef3a5937df6107c2"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.618966    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7qjxk"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.623990    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7qjxk"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.636035    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cbw6w"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.639767    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.645577    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xfnlv"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.648495    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qvd2r"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.651340    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qvd2r"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.658922    4824 scope.go:117] "RemoveContainer" containerID="818ce7d2540b4e6464a5ff26be6cb503b5edb5c2cddfd5c3d68aab18c0efe0fa"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.662652    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9rdhx"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.670933    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9rdhx"]
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.686652    4824 scope.go:117] "RemoveContainer" containerID="1419fdfd5269cfff5e11ab0c006f882559be5f169910c27a72809be04a43be39"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.703479    4824 scope.go:117] "RemoveContainer" containerID="7317b7210034e3aa2ec35c18a34feca5f748280edc9c30e6422e019e9b037d9f"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.727761    4824 scope.go:117] "RemoveContainer" containerID="ef1771a84cf61915d4b0d975f9ee8872b5cd82196ad2083f1dd983ba605b8a06"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.761947    4824 scope.go:117] "RemoveContainer" containerID="40981f624a90205622eab9f1ca631923b8deed71d17240a64a0f9f2432fa979b"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.779076    4824 scope.go:117] "RemoveContainer" containerID="7742bb669ebb371304c069b943f41fb6f0310aba705fbfcb09c46448be25cfa2"
Jan 22 09:48:18 crc kubenswrapper[4824]: I0122 09:48:18.796649    4824 scope.go:117] "RemoveContainer" containerID="11133aa1bdf4f62b5e6e2961b375b90475d985dd21a6ecbc90f2326ffca0d637"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.247819    4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7qjxk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body=
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.247914    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7qjxk" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.410429    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" path="/var/lib/kubelet/pods/05d38c4c-ca18-478a-bcd5-01800ded772a/volumes"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.411032    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072ef9f4-8754-45c5-9600-accb760d786c" path="/var/lib/kubelet/pods/072ef9f4-8754-45c5-9600-accb760d786c/volumes"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.411643    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" path="/var/lib/kubelet/pods/2463610b-364e-4154-83ee-b6c00a6c6da5/volumes"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.412154    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b498256-7f0d-4941-998e-e670f5941cdb" path="/var/lib/kubelet/pods/2b498256-7f0d-4941-998e-e670f5941cdb/volumes"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.412794    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88bd6885-a443-4d1f-b06e-5dd0f3107dab" path="/var/lib/kubelet/pods/88bd6885-a443-4d1f-b06e-5dd0f3107dab/volumes"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.413511    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2678ef-e863-4181-a497-b98b7fff0cce" path="/var/lib/kubelet/pods/8d2678ef-e863-4181-a497-b98b7fff0cce/volumes"
Jan 22 09:48:19 crc kubenswrapper[4824]: I0122 09:48:19.593304    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w" event={"ID":"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6","Type":"ContainerStarted","Data":"73d5104cfbfef2b1d7ce718de18333f0ff3358b55bdfc876ee0ad27ec6ae54a0"}
Jan 22 09:48:23 crc kubenswrapper[4824]: I0122 09:48:23.625500    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w" event={"ID":"e2af3857-54b3-4f97-9b57-c8b2afdcaaa6","Type":"ContainerStarted","Data":"54b75e7e1d39f64e3689ee9229d1b2d597f2560a88cf314fc46d740e9b2350c4"}
Jan 22 09:48:23 crc kubenswrapper[4824]: I0122 09:48:23.626943    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:23 crc kubenswrapper[4824]: I0122 09:48:23.650945    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w" podStartSLOduration=6.650929586 podStartE2EDuration="6.650929586s" podCreationTimestamp="2026-01-22 09:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:48:23.649308121 +0000 UTC m=+336.552785114" watchObservedRunningTime="2026-01-22 09:48:23.650929586 +0000 UTC m=+336.554406579"
Jan 22 09:48:38 crc kubenswrapper[4824]: I0122 09:48:38.193086    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-cbw6w"
Jan 22 09:48:38 crc kubenswrapper[4824]: I0122 09:48:38.250799    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbdgw"]
Jan 22 09:49:00 crc kubenswrapper[4824]: I0122 09:49:00.333310    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:49:00 crc kubenswrapper[4824]: I0122 09:49:00.333902    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.478733    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ksxrb"]
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.480089    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="extract-content"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.480258    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="extract-content"
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.480525    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="registry-server"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.480691    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="registry-server"
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.480836    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="extract-utilities"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.483407    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="extract-utilities"
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.483618    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="registry-server"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.483820    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="registry-server"
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.483986    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="extract-content"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.484128    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="extract-content"
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.484275    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.484441    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator"
Jan 22 09:49:02 crc kubenswrapper[4824]: E0122 09:49:02.484653    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="extract-utilities"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.485269    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="extract-utilities"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.485858    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="05d38c4c-ca18-478a-bcd5-01800ded772a" containerName="registry-server"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.486034    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="072ef9f4-8754-45c5-9600-accb760d786c" containerName="registry-server"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.486183    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2463610b-364e-4154-83ee-b6c00a6c6da5" containerName="marketplace-operator"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.487706    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.490836    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ksxrb"]
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.501126    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.518800    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b28r2\" (UniqueName: \"kubernetes.io/projected/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-kube-api-access-b28r2\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.518899    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-utilities\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.518918    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-catalog-content\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.620011    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b28r2\" (UniqueName: \"kubernetes.io/projected/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-kube-api-access-b28r2\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.620622    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-utilities\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.620650    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-catalog-content\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.621420    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-catalog-content\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.622869    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-utilities\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.645560    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b28r2\" (UniqueName: \"kubernetes.io/projected/03cdd9c2-a33d-4b26-adee-57d691ac5eb9-kube-api-access-b28r2\") pod \"redhat-operators-ksxrb\" (UID: \"03cdd9c2-a33d-4b26-adee-57d691ac5eb9\") " pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.663165    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c4wsw"]
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.664696    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.666718    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.675784    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4wsw"]
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.721747    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ac9b68-23da-4599-b9bc-84546dbd1c52-catalog-content\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.722077    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ac9b68-23da-4599-b9bc-84546dbd1c52-utilities\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.722227    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw7fj\" (UniqueName: \"kubernetes.io/projected/33ac9b68-23da-4599-b9bc-84546dbd1c52-kube-api-access-jw7fj\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.814102    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.824150    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ac9b68-23da-4599-b9bc-84546dbd1c52-catalog-content\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.824233    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ac9b68-23da-4599-b9bc-84546dbd1c52-utilities\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.824273    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw7fj\" (UniqueName: \"kubernetes.io/projected/33ac9b68-23da-4599-b9bc-84546dbd1c52-kube-api-access-jw7fj\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.825082    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33ac9b68-23da-4599-b9bc-84546dbd1c52-utilities\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.825239    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33ac9b68-23da-4599-b9bc-84546dbd1c52-catalog-content\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.851114    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw7fj\" (UniqueName: \"kubernetes.io/projected/33ac9b68-23da-4599-b9bc-84546dbd1c52-kube-api-access-jw7fj\") pod \"redhat-marketplace-c4wsw\" (UID: \"33ac9b68-23da-4599-b9bc-84546dbd1c52\") " pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:02 crc kubenswrapper[4824]: I0122 09:49:02.995146    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.212967    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ksxrb"]
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.297490    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" podUID="9633603d-c8b9-4f94-a3c5-3d882308a586" containerName="registry" containerID="cri-o://8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd" gracePeriod=30
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.411227    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c4wsw"]
Jan 22 09:49:03 crc kubenswrapper[4824]: W0122 09:49:03.467714    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33ac9b68_23da_4599_b9bc_84546dbd1c52.slice/crio-fb7a607a6a0203556d4d66516d14e4bb60062feddac2bd2cae9064683cb18d8f WatchSource:0}: Error finding container fb7a607a6a0203556d4d66516d14e4bb60062feddac2bd2cae9064683cb18d8f: Status 404 returned error can't find the container with id fb7a607a6a0203556d4d66516d14e4bb60062feddac2bd2cae9064683cb18d8f
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.703649    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.735316    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9633603d-c8b9-4f94-a3c5-3d882308a586-ca-trust-extracted\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.735436    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9633603d-c8b9-4f94-a3c5-3d882308a586-installation-pull-secrets\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.735689    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.735800    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-trusted-ca\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.736426    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.736524    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-bound-sa-token\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.736895    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-certificates\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.736926    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tv7t\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-kube-api-access-7tv7t\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.736943    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-tls\") pod \"9633603d-c8b9-4f94-a3c5-3d882308a586\" (UID: \"9633603d-c8b9-4f94-a3c5-3d882308a586\") "
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.737448    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-trusted-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.737575    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.741677    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-kube-api-access-7tv7t" (OuterVolumeSpecName: "kube-api-access-7tv7t") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "kube-api-access-7tv7t". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.741976    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.742708    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9633603d-c8b9-4f94-a3c5-3d882308a586-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.742987    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.743504    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.753307    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9633603d-c8b9-4f94-a3c5-3d882308a586-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9633603d-c8b9-4f94-a3c5-3d882308a586" (UID: "9633603d-c8b9-4f94-a3c5-3d882308a586"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.838174    4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-bound-sa-token\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.838240    4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-certificates\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.838255    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tv7t\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-kube-api-access-7tv7t\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.838267    4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9633603d-c8b9-4f94-a3c5-3d882308a586-registry-tls\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.838281    4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9633603d-c8b9-4f94-a3c5-3d882308a586-ca-trust-extracted\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.838294    4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9633603d-c8b9-4f94-a3c5-3d882308a586-installation-pull-secrets\") on node \"crc\" DevicePath \"\""
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.856345    4824 generic.go:334] "Generic (PLEG): container finished" podID="03cdd9c2-a33d-4b26-adee-57d691ac5eb9" containerID="0fea5dd59d6853c74683fcf5207682dbc1deaeeddd9784522d4ca3ddea9e152b" exitCode=0
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.856424    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksxrb" event={"ID":"03cdd9c2-a33d-4b26-adee-57d691ac5eb9","Type":"ContainerDied","Data":"0fea5dd59d6853c74683fcf5207682dbc1deaeeddd9784522d4ca3ddea9e152b"}
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.856492    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksxrb" event={"ID":"03cdd9c2-a33d-4b26-adee-57d691ac5eb9","Type":"ContainerStarted","Data":"657302e920536d87771e4f8ebf2e22c1e3da447005d540d8ea9936719d3f1891"}
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.860002    4824 generic.go:334] "Generic (PLEG): container finished" podID="33ac9b68-23da-4599-b9bc-84546dbd1c52" containerID="c5beb0cc8e44c30725a1b6e7746df4afb26b3b2a246869737aee3df5668ab39c" exitCode=0
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.860061    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4wsw" event={"ID":"33ac9b68-23da-4599-b9bc-84546dbd1c52","Type":"ContainerDied","Data":"c5beb0cc8e44c30725a1b6e7746df4afb26b3b2a246869737aee3df5668ab39c"}
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.860086    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4wsw" event={"ID":"33ac9b68-23da-4599-b9bc-84546dbd1c52","Type":"ContainerStarted","Data":"fb7a607a6a0203556d4d66516d14e4bb60062feddac2bd2cae9064683cb18d8f"}
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.863137    4824 generic.go:334] "Generic (PLEG): container finished" podID="9633603d-c8b9-4f94-a3c5-3d882308a586" containerID="8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd" exitCode=0
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.863176    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" event={"ID":"9633603d-c8b9-4f94-a3c5-3d882308a586","Type":"ContainerDied","Data":"8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd"}
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.863229    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw"
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.863244    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zbdgw" event={"ID":"9633603d-c8b9-4f94-a3c5-3d882308a586","Type":"ContainerDied","Data":"73438183f6f933033b7002f1b6135147cae0411ee1dd6fe5d90fc317f68177d1"}
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.863268    4824 scope.go:117] "RemoveContainer" containerID="8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd"
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.881918    4824 scope.go:117] "RemoveContainer" containerID="8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd"
Jan 22 09:49:03 crc kubenswrapper[4824]: E0122 09:49:03.882467    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd\": container with ID starting with 8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd not found: ID does not exist" containerID="8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd"
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.882506    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd"} err="failed to get container status \"8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd\": rpc error: code = NotFound desc = could not find container \"8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd\": container with ID starting with 8dba216b4a5d8893667d7d71a7c642e0097ac76c0ef08b3bdd08cf14fb2577bd not found: ID does not exist"
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.914567    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbdgw"]
Jan 22 09:49:03 crc kubenswrapper[4824]: I0122 09:49:03.920265    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zbdgw"]
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.865424    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pb5fd"]
Jan 22 09:49:04 crc kubenswrapper[4824]: E0122 09:49:04.865980    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9633603d-c8b9-4f94-a3c5-3d882308a586" containerName="registry"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.865994    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9633603d-c8b9-4f94-a3c5-3d882308a586" containerName="registry"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.866103    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9633603d-c8b9-4f94-a3c5-3d882308a586" containerName="registry"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.866936    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.869742    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.871719    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4wsw" event={"ID":"33ac9b68-23da-4599-b9bc-84546dbd1c52","Type":"ContainerStarted","Data":"1e4fcc4b649c0e0ee1a3b2a9e305ef65e56af6a63b9b2e24772ed725e1f36df8"}
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.876463    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksxrb" event={"ID":"03cdd9c2-a33d-4b26-adee-57d691ac5eb9","Type":"ContainerStarted","Data":"33604cd27c4730616694a10d1108ee01efc44c9ce49be22143a0ddf0119cc013"}
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.886230    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pb5fd"]
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.960414    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-catalog-content\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.960471    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-utilities\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:04 crc kubenswrapper[4824]: I0122 09:49:04.960516    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxhqm\" (UniqueName: \"kubernetes.io/projected/72d1dd96-6496-4073-b2db-de3451546d7c-kube-api-access-gxhqm\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.062529    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-catalog-content\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.062659    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-utilities\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.062741    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxhqm\" (UniqueName: \"kubernetes.io/projected/72d1dd96-6496-4073-b2db-de3451546d7c-kube-api-access-gxhqm\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.063191    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-catalog-content\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.063507    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9ns86"]
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.063559    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-utilities\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.064521    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.067259    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.084398    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxhqm\" (UniqueName: \"kubernetes.io/projected/72d1dd96-6496-4073-b2db-de3451546d7c-kube-api-access-gxhqm\") pod \"certified-operators-pb5fd\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") " pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.092003    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9ns86"]
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.164080    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed2abbd-725c-4738-9e66-d7143918cb30-catalog-content\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.164334    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vck6x\" (UniqueName: \"kubernetes.io/projected/1ed2abbd-725c-4738-9e66-d7143918cb30-kube-api-access-vck6x\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.164558    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed2abbd-725c-4738-9e66-d7143918cb30-utilities\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.181220    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.267533    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed2abbd-725c-4738-9e66-d7143918cb30-catalog-content\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.267825    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vck6x\" (UniqueName: \"kubernetes.io/projected/1ed2abbd-725c-4738-9e66-d7143918cb30-kube-api-access-vck6x\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.267902    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed2abbd-725c-4738-9e66-d7143918cb30-utilities\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.268659    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ed2abbd-725c-4738-9e66-d7143918cb30-utilities\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.268827    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ed2abbd-725c-4738-9e66-d7143918cb30-catalog-content\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.283713    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vck6x\" (UniqueName: \"kubernetes.io/projected/1ed2abbd-725c-4738-9e66-d7143918cb30-kube-api-access-vck6x\") pod \"community-operators-9ns86\" (UID: \"1ed2abbd-725c-4738-9e66-d7143918cb30\") " pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.385311    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.411706    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9633603d-c8b9-4f94-a3c5-3d882308a586" path="/var/lib/kubelet/pods/9633603d-c8b9-4f94-a3c5-3d882308a586/volumes"
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.578608    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pb5fd"]
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.755988    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9ns86"]
Jan 22 09:49:05 crc kubenswrapper[4824]: W0122 09:49:05.795397    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ed2abbd_725c_4738_9e66_d7143918cb30.slice/crio-fab9f69d9feb810b79a70bc7d89663bbcc62df2cb2e9a0dc2246dd9dc81ac7bd WatchSource:0}: Error finding container fab9f69d9feb810b79a70bc7d89663bbcc62df2cb2e9a0dc2246dd9dc81ac7bd: Status 404 returned error can't find the container with id fab9f69d9feb810b79a70bc7d89663bbcc62df2cb2e9a0dc2246dd9dc81ac7bd
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.884607    4824 generic.go:334] "Generic (PLEG): container finished" podID="33ac9b68-23da-4599-b9bc-84546dbd1c52" containerID="1e4fcc4b649c0e0ee1a3b2a9e305ef65e56af6a63b9b2e24772ed725e1f36df8" exitCode=0
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.884690    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4wsw" event={"ID":"33ac9b68-23da-4599-b9bc-84546dbd1c52","Type":"ContainerDied","Data":"1e4fcc4b649c0e0ee1a3b2a9e305ef65e56af6a63b9b2e24772ed725e1f36df8"}
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.887231    4824 generic.go:334] "Generic (PLEG): container finished" podID="72d1dd96-6496-4073-b2db-de3451546d7c" containerID="1820fa629419cbc44acc407648aab895ae74917da5e988f704c2af16b621d95d" exitCode=0
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.887293    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pb5fd" event={"ID":"72d1dd96-6496-4073-b2db-de3451546d7c","Type":"ContainerDied","Data":"1820fa629419cbc44acc407648aab895ae74917da5e988f704c2af16b621d95d"}
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.887337    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pb5fd" event={"ID":"72d1dd96-6496-4073-b2db-de3451546d7c","Type":"ContainerStarted","Data":"c8386719eeeeeb0778791f9fc7b212358c0bf5abe62fb48bdd10c32fe9f29403"}
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.888792    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ns86" event={"ID":"1ed2abbd-725c-4738-9e66-d7143918cb30","Type":"ContainerStarted","Data":"fab9f69d9feb810b79a70bc7d89663bbcc62df2cb2e9a0dc2246dd9dc81ac7bd"}
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.892937    4824 generic.go:334] "Generic (PLEG): container finished" podID="03cdd9c2-a33d-4b26-adee-57d691ac5eb9" containerID="33604cd27c4730616694a10d1108ee01efc44c9ce49be22143a0ddf0119cc013" exitCode=0
Jan 22 09:49:05 crc kubenswrapper[4824]: I0122 09:49:05.892975    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksxrb" event={"ID":"03cdd9c2-a33d-4b26-adee-57d691ac5eb9","Type":"ContainerDied","Data":"33604cd27c4730616694a10d1108ee01efc44c9ce49be22143a0ddf0119cc013"}
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.899524    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ksxrb" event={"ID":"03cdd9c2-a33d-4b26-adee-57d691ac5eb9","Type":"ContainerStarted","Data":"9d0f966803a75489d29f8e510bf3bc88adc4ce3e6800dea56e5f10985492d19e"}
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.902351    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c4wsw" event={"ID":"33ac9b68-23da-4599-b9bc-84546dbd1c52","Type":"ContainerStarted","Data":"b818a5b3258839e6dcd35d7b39fe6717146da655216452215cb8f2ad08d15e95"}
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.905321    4824 generic.go:334] "Generic (PLEG): container finished" podID="72d1dd96-6496-4073-b2db-de3451546d7c" containerID="cfb02a68f1f8b38b043e4204a819a352d1036210349ce3e0e252c02259e68c99" exitCode=0
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.905506    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pb5fd" event={"ID":"72d1dd96-6496-4073-b2db-de3451546d7c","Type":"ContainerDied","Data":"cfb02a68f1f8b38b043e4204a819a352d1036210349ce3e0e252c02259e68c99"}
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.907736    4824 generic.go:334] "Generic (PLEG): container finished" podID="1ed2abbd-725c-4738-9e66-d7143918cb30" containerID="acacb6d5438c84de8250faa33e0f99ca19a6a9946e159fd033d6ea562c8a9fad" exitCode=0
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.907827    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ns86" event={"ID":"1ed2abbd-725c-4738-9e66-d7143918cb30","Type":"ContainerDied","Data":"acacb6d5438c84de8250faa33e0f99ca19a6a9946e159fd033d6ea562c8a9fad"}
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.924438    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ksxrb" podStartSLOduration=2.459120841 podStartE2EDuration="4.92442417s" podCreationTimestamp="2026-01-22 09:49:02 +0000 UTC" firstStartedPulling="2026-01-22 09:49:03.85778013 +0000 UTC m=+376.761257123" lastFinishedPulling="2026-01-22 09:49:06.323083449 +0000 UTC m=+379.226560452" observedRunningTime="2026-01-22 09:49:06.921837209 +0000 UTC m=+379.825314192" watchObservedRunningTime="2026-01-22 09:49:06.92442417 +0000 UTC m=+379.827901163"
Jan 22 09:49:06 crc kubenswrapper[4824]: I0122 09:49:06.955681    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c4wsw" podStartSLOduration=2.549768838 podStartE2EDuration="4.9556672s" podCreationTimestamp="2026-01-22 09:49:02 +0000 UTC" firstStartedPulling="2026-01-22 09:49:03.861528122 +0000 UTC m=+376.765005115" lastFinishedPulling="2026-01-22 09:49:06.267426444 +0000 UTC m=+379.170903477" observedRunningTime="2026-01-22 09:49:06.953145921 +0000 UTC m=+379.856622914" watchObservedRunningTime="2026-01-22 09:49:06.9556672 +0000 UTC m=+379.859144183"
Jan 22 09:49:07 crc kubenswrapper[4824]: I0122 09:49:07.913733    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ns86" event={"ID":"1ed2abbd-725c-4738-9e66-d7143918cb30","Type":"ContainerStarted","Data":"1b5341147fbba223feec7000aacd1a87306511e1b0b0654b89d1b2b752ce13d6"}
Jan 22 09:49:07 crc kubenswrapper[4824]: I0122 09:49:07.916511    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pb5fd" event={"ID":"72d1dd96-6496-4073-b2db-de3451546d7c","Type":"ContainerStarted","Data":"fca4a3d5e4e5ed5496d18732969eb01f42d50954ed6aaf181cc79205888870b2"}
Jan 22 09:49:07 crc kubenswrapper[4824]: I0122 09:49:07.947764    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pb5fd" podStartSLOduration=2.454890764 podStartE2EDuration="3.947743653s" podCreationTimestamp="2026-01-22 09:49:04 +0000 UTC" firstStartedPulling="2026-01-22 09:49:05.890749645 +0000 UTC m=+378.794226638" lastFinishedPulling="2026-01-22 09:49:07.383602534 +0000 UTC m=+380.287079527" observedRunningTime="2026-01-22 09:49:07.943904169 +0000 UTC m=+380.847381212" watchObservedRunningTime="2026-01-22 09:49:07.947743653 +0000 UTC m=+380.851220656"
Jan 22 09:49:08 crc kubenswrapper[4824]: I0122 09:49:08.923250    4824 generic.go:334] "Generic (PLEG): container finished" podID="1ed2abbd-725c-4738-9e66-d7143918cb30" containerID="1b5341147fbba223feec7000aacd1a87306511e1b0b0654b89d1b2b752ce13d6" exitCode=0
Jan 22 09:49:08 crc kubenswrapper[4824]: I0122 09:49:08.923374    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ns86" event={"ID":"1ed2abbd-725c-4738-9e66-d7143918cb30","Type":"ContainerDied","Data":"1b5341147fbba223feec7000aacd1a87306511e1b0b0654b89d1b2b752ce13d6"}
Jan 22 09:49:09 crc kubenswrapper[4824]: I0122 09:49:09.930651    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ns86" event={"ID":"1ed2abbd-725c-4738-9e66-d7143918cb30","Type":"ContainerStarted","Data":"5367b2abee0b6df49cd2e3a96de7c5143c6f38bbd96a58e7477d86ffdf0fde6b"}
Jan 22 09:49:09 crc kubenswrapper[4824]: I0122 09:49:09.950299    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9ns86" podStartSLOduration=2.4566831430000002 podStartE2EDuration="4.950278101s" podCreationTimestamp="2026-01-22 09:49:05 +0000 UTC" firstStartedPulling="2026-01-22 09:49:06.909756971 +0000 UTC m=+379.813233964" lastFinishedPulling="2026-01-22 09:49:09.403351919 +0000 UTC m=+382.306828922" observedRunningTime="2026-01-22 09:49:09.946811746 +0000 UTC m=+382.850288749" watchObservedRunningTime="2026-01-22 09:49:09.950278101 +0000 UTC m=+382.853755094"
Jan 22 09:49:12 crc kubenswrapper[4824]: I0122 09:49:12.815228    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:12 crc kubenswrapper[4824]: I0122 09:49:12.815667    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:12 crc kubenswrapper[4824]: I0122 09:49:12.880905    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:12 crc kubenswrapper[4824]: I0122 09:49:12.992039    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ksxrb"
Jan 22 09:49:12 crc kubenswrapper[4824]: I0122 09:49:12.995295    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:12 crc kubenswrapper[4824]: I0122 09:49:12.995343    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:13 crc kubenswrapper[4824]: I0122 09:49:13.037093    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:14 crc kubenswrapper[4824]: I0122 09:49:14.020033    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c4wsw"
Jan 22 09:49:15 crc kubenswrapper[4824]: I0122 09:49:15.181955    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:15 crc kubenswrapper[4824]: I0122 09:49:15.182016    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:15 crc kubenswrapper[4824]: I0122 09:49:15.245808    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:15 crc kubenswrapper[4824]: I0122 09:49:15.386306    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:15 crc kubenswrapper[4824]: I0122 09:49:15.386456    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:15 crc kubenswrapper[4824]: I0122 09:49:15.426527    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:16 crc kubenswrapper[4824]: I0122 09:49:16.032695    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9ns86"
Jan 22 09:49:16 crc kubenswrapper[4824]: I0122 09:49:16.033702    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:49:30 crc kubenswrapper[4824]: I0122 09:49:30.332806    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:49:30 crc kubenswrapper[4824]: I0122 09:49:30.333445    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:50:00 crc kubenswrapper[4824]: I0122 09:50:00.333327    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:50:00 crc kubenswrapper[4824]: I0122 09:50:00.334058    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:50:00 crc kubenswrapper[4824]: I0122 09:50:00.334133    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:50:00 crc kubenswrapper[4824]: I0122 09:50:00.335084    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b8ff90b63f9c4bc3e35eb99a32bf8aff84f6fad5953f3a33158e75564461092"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 09:50:00 crc kubenswrapper[4824]: I0122 09:50:00.335183    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://7b8ff90b63f9c4bc3e35eb99a32bf8aff84f6fad5953f3a33158e75564461092" gracePeriod=600
Jan 22 09:50:01 crc kubenswrapper[4824]: I0122 09:50:01.214955    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="7b8ff90b63f9c4bc3e35eb99a32bf8aff84f6fad5953f3a33158e75564461092" exitCode=0
Jan 22 09:50:01 crc kubenswrapper[4824]: I0122 09:50:01.215008    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"7b8ff90b63f9c4bc3e35eb99a32bf8aff84f6fad5953f3a33158e75564461092"}
Jan 22 09:50:01 crc kubenswrapper[4824]: I0122 09:50:01.215285    4824 scope.go:117] "RemoveContainer" containerID="ae57c99ec8877c5d1d5ff7decad372fa4272018f4fefd6b18519045843d0a784"
Jan 22 09:50:02 crc kubenswrapper[4824]: I0122 09:50:02.227502    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"825ee924b14e8fa29e4a235d69b6ea74e054515fdadb5e9acc65891a65f1dced"}
Jan 22 09:52:30 crc kubenswrapper[4824]: I0122 09:52:30.333474    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:52:30 crc kubenswrapper[4824]: I0122 09:52:30.334274    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:53:00 crc kubenswrapper[4824]: I0122 09:53:00.332942    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:53:00 crc kubenswrapper[4824]: I0122 09:53:00.333476    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:53:30 crc kubenswrapper[4824]: I0122 09:53:30.333271    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:53:30 crc kubenswrapper[4824]: I0122 09:53:30.333742    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:53:30 crc kubenswrapper[4824]: I0122 09:53:30.333794    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:53:30 crc kubenswrapper[4824]: I0122 09:53:30.334773    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"825ee924b14e8fa29e4a235d69b6ea74e054515fdadb5e9acc65891a65f1dced"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 09:53:30 crc kubenswrapper[4824]: I0122 09:53:30.334904    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://825ee924b14e8fa29e4a235d69b6ea74e054515fdadb5e9acc65891a65f1dced" gracePeriod=600
Jan 22 09:53:31 crc kubenswrapper[4824]: I0122 09:53:31.441437    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="825ee924b14e8fa29e4a235d69b6ea74e054515fdadb5e9acc65891a65f1dced" exitCode=0
Jan 22 09:53:31 crc kubenswrapper[4824]: I0122 09:53:31.441648    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"825ee924b14e8fa29e4a235d69b6ea74e054515fdadb5e9acc65891a65f1dced"}
Jan 22 09:53:31 crc kubenswrapper[4824]: I0122 09:53:31.442520    4824 scope.go:117] "RemoveContainer" containerID="7b8ff90b63f9c4bc3e35eb99a32bf8aff84f6fad5953f3a33158e75564461092"
Jan 22 09:53:32 crc kubenswrapper[4824]: I0122 09:53:32.451145    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"897b49c4db7fcfddc6a831eea83804b055a91a5870ae0d9b09b9c435efb91273"}
Jan 22 09:55:29 crc kubenswrapper[4824]: I0122 09:55:29.886907    4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt"
Jan 22 09:56:00 crc kubenswrapper[4824]: I0122 09:56:00.333103    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:56:00 crc kubenswrapper[4824]: I0122 09:56:00.334872    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.114752    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8vt4"]
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.115719    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-controller" containerID="cri-o://430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.115832    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.115867    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-node" containerID="cri-o://466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.115831    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="northd" containerID="cri-o://070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.115897    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-acl-logging" containerID="cri-o://fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.116058    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="nbdb" containerID="cri-o://37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.116065    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="sbdb" containerID="cri-o://20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.210525    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller" containerID="cri-o://0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5" gracePeriod=30
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.442779    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/3.log"
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.445506    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovn-acl-logging/0.log"
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446052    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovn-controller/0.log"
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446490    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524" exitCode=0
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446516    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c" exitCode=0
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446525    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b" exitCode=143
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446537    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20" exitCode=143
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446576    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"}
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446642    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"}
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446662    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"}
Jan 22 09:56:26 crc kubenswrapper[4824]: I0122 09:56:26.446681    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.086195    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/3.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.088914    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovn-acl-logging/0.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.089576    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovn-controller/0.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.090199    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137267    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-log-socket\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137329    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-openvswitch\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137386    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-netns\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137415    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-systemd-units\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137426    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-log-socket" (OuterVolumeSpecName: "log-socket") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137459    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-var-lib-openvswitch\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137478    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137496    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-ovn-kubernetes\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137502    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137525    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137525    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-kubelet\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137571    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-etc-openvswitch\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137581    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137598    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovn-node-metrics-cert\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137627    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137653    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-script-lib\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137663    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137671    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137682    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137716    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137697    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-netd\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137784    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-config\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137818    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pq4s\" (UniqueName: \"kubernetes.io/projected/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-kube-api-access-5pq4s\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137864    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-slash\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137888    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-bin\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137917    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-env-overrides\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137939    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-ovn\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137964    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-systemd\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.137984    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-node-log\") pod \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\" (UID: \"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb\") "
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138096    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138126    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138148    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138340    4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-log-socket\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138359    4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-openvswitch\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138391    4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-netns\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138403    4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-systemd-units\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138416    4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-var-lib-openvswitch\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138428    4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138439    4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-kubelet\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138451    4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-etc-openvswitch\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138462    4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-script-lib\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138474    4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138486    4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-netd\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138498    4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-cni-bin\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138360    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138506    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-node-log" (OuterVolumeSpecName: "node-log") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138484    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-slash" (OuterVolumeSpecName: "host-slash") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138458    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.138784    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.146639    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-kube-api-access-5pq4s" (OuterVolumeSpecName: "kube-api-access-5pq4s") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "kube-api-access-5pq4s". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.157938    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dwgg4"]
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158135    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kubecfg-setup"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158147    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kubecfg-setup"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158156    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158162    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158174    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158180    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158189    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158194    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158201    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-ovn-metrics"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158207    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-ovn-metrics"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158213    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="nbdb"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158219    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="nbdb"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158225    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-node"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158232    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-node"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158241    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="sbdb"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158248    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="sbdb"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158255    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158260    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158269    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="northd"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158275    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="northd"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158281    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158287    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158302    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-acl-logging"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158313    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-acl-logging"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158455    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="sbdb"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158475    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158484    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158492    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="nbdb"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158499    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158504    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158513    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158520    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-node"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158527    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="kube-rbac-proxy-ovn-metrics"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158534    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158540    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="northd"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158548    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovn-acl-logging"
Jan 22 09:56:27 crc kubenswrapper[4824]: E0122 09:56:27.158631    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.158638    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerName="ovnkube-controller"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.160114    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.165988    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.171726    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" (UID: "40c9681a-fa96-4b9f-8ff5-2d81bbf190fb"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.239997    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-cni-bin\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240039    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-etc-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240065    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-ovn\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240086    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-systemd\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240107    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-node-log\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240129    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9qd7\" (UniqueName: \"kubernetes.io/projected/59c1deb5-4a8b-4333-8a39-577a456beae1-kube-api-access-g9qd7\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240150    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-run-netns\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240278    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-ovnkube-config\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240340    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240419    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-log-socket\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240493    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240533    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-ovnkube-script-lib\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240603    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59c1deb5-4a8b-4333-8a39-577a456beae1-ovn-node-metrics-cert\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240636    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-var-lib-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240659    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-run-ovn-kubernetes\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240680    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-cni-netd\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240771    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-slash\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240845    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-systemd-units\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240867    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-kubelet\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240884    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-env-overrides\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240961    4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240975    4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-ovnkube-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240984    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pq4s\" (UniqueName: \"kubernetes.io/projected/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-kube-api-access-5pq4s\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.240994    4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-host-slash\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.241016    4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-env-overrides\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.241039    4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-ovn\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.241053    4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-run-systemd\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.241065    4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb-node-log\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.341917    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-var-lib-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342015    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-var-lib-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342038    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-run-ovn-kubernetes\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342083    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-cni-netd\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342116    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-slash\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342142    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-run-ovn-kubernetes\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342170    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-cni-netd\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342184    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-systemd-units\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342226    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-slash\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342161    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-systemd-units\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342253    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-kubelet\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342287    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-env-overrides\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342304    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-cni-bin\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342384    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-cni-bin\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342397    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-kubelet\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.342493    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-etc-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343205    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-env-overrides\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343246    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-etc-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343281    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-ovn\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343296    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-systemd\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343312    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-node-log\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343350    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-ovn\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343415    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9qd7\" (UniqueName: \"kubernetes.io/projected/59c1deb5-4a8b-4333-8a39-577a456beae1-kube-api-access-g9qd7\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343452    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-systemd\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343477    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-node-log\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343570    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-run-netns\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343800    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-run-netns\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343848    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-ovnkube-config\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.343871    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344071    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-run-openvswitch\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344410    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-ovnkube-config\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344507    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-log-socket\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344578    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-log-socket\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344607    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344627    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-ovnkube-script-lib\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344682    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/59c1deb5-4a8b-4333-8a39-577a456beae1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.345286    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/59c1deb5-4a8b-4333-8a39-577a456beae1-ovnkube-script-lib\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.344745    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59c1deb5-4a8b-4333-8a39-577a456beae1-ovn-node-metrics-cert\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.349870    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/59c1deb5-4a8b-4333-8a39-577a456beae1-ovn-node-metrics-cert\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.366009    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9qd7\" (UniqueName: \"kubernetes.io/projected/59c1deb5-4a8b-4333-8a39-577a456beae1-kube-api-access-g9qd7\") pod \"ovnkube-node-dwgg4\" (UID: \"59c1deb5-4a8b-4333-8a39-577a456beae1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.453763    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/2.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.454227    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/1.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.454291    4824 generic.go:334] "Generic (PLEG): container finished" podID="00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c" containerID="158628e706616c42e59547ac0e862beff096705bca967864fa84347bcd7b8aa5" exitCode=2
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.454393    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerDied","Data":"158628e706616c42e59547ac0e862beff096705bca967864fa84347bcd7b8aa5"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.454452    4824 scope.go:117] "RemoveContainer" containerID="814ef031a1a3094d44cc4566cc78b275fe9615c5a8d71f588f3550447adc8914"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.455181    4824 scope.go:117] "RemoveContainer" containerID="158628e706616c42e59547ac0e862beff096705bca967864fa84347bcd7b8aa5"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.456591    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovnkube-controller/3.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.469586    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovn-acl-logging/0.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470342    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s8vt4_40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/ovn-controller/0.log"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470873    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5" exitCode=0
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470901    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d" exitCode=0
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470909    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84" exitCode=0
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470919    4824 generic.go:334] "Generic (PLEG): container finished" podID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7" exitCode=0
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470943    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470973    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.470987    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.471000    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.471014    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4" event={"ID":"40c9681a-fa96-4b9f-8ff5-2d81bbf190fb","Type":"ContainerDied","Data":"24ee1eb98143886fb2956ccf86fee1e5aab73aa59d6d3fbd9bba7ab3b071fc3d"}
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.471114    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s8vt4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.480782    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.509478    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8vt4"]
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.522069    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s8vt4"]
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.774239    4824 scope.go:117] "RemoveContainer" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.797165    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.819986    4824 scope.go:117] "RemoveContainer" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.837013    4824 scope.go:117] "RemoveContainer" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.867483    4824 scope.go:117] "RemoveContainer" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.885474    4824 scope.go:117] "RemoveContainer" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.903077    4824 scope.go:117] "RemoveContainer" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.917089    4824 scope.go:117] "RemoveContainer" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.933641    4824 scope.go:117] "RemoveContainer" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"
Jan 22 09:56:27 crc kubenswrapper[4824]: I0122 09:56:27.991207    4824 scope.go:117] "RemoveContainer" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.008599    4824 scope.go:117] "RemoveContainer" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.009186    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": container with ID starting with 0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5 not found: ID does not exist" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.009371    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"} err="failed to get container status \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": rpc error: code = NotFound desc = could not find container \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": container with ID starting with 0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.009406    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.009920    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": container with ID starting with f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646 not found: ID does not exist" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.009959    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"} err="failed to get container status \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": rpc error: code = NotFound desc = could not find container \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": container with ID starting with f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.009988    4824 scope.go:117] "RemoveContainer" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.010420    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": container with ID starting with 20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d not found: ID does not exist" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.010481    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"} err="failed to get container status \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": rpc error: code = NotFound desc = could not find container \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": container with ID starting with 20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.010516    4824 scope.go:117] "RemoveContainer" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.010946    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": container with ID starting with 37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84 not found: ID does not exist" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.010978    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"} err="failed to get container status \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": rpc error: code = NotFound desc = could not find container \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": container with ID starting with 37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.011002    4824 scope.go:117] "RemoveContainer" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.011524    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": container with ID starting with 070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7 not found: ID does not exist" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.011553    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"} err="failed to get container status \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": rpc error: code = NotFound desc = could not find container \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": container with ID starting with 070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.011572    4824 scope.go:117] "RemoveContainer" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.011872    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": container with ID starting with 0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524 not found: ID does not exist" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.011903    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"} err="failed to get container status \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": rpc error: code = NotFound desc = could not find container \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": container with ID starting with 0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.011923    4824 scope.go:117] "RemoveContainer" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.012321    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": container with ID starting with 466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c not found: ID does not exist" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.012346    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"} err="failed to get container status \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": rpc error: code = NotFound desc = could not find container \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": container with ID starting with 466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.012386    4824 scope.go:117] "RemoveContainer" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.012832    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": container with ID starting with fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b not found: ID does not exist" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.012867    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"} err="failed to get container status \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": rpc error: code = NotFound desc = could not find container \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": container with ID starting with fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.012888    4824 scope.go:117] "RemoveContainer" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.013201    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": container with ID starting with 430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20 not found: ID does not exist" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.013247    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"} err="failed to get container status \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": rpc error: code = NotFound desc = could not find container \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": container with ID starting with 430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.013277    4824 scope.go:117] "RemoveContainer" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"
Jan 22 09:56:28 crc kubenswrapper[4824]: E0122 09:56:28.013862    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": container with ID starting with e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35 not found: ID does not exist" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.013899    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"} err="failed to get container status \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": rpc error: code = NotFound desc = could not find container \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": container with ID starting with e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.013937    4824 scope.go:117] "RemoveContainer" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.014325    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"} err="failed to get container status \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": rpc error: code = NotFound desc = could not find container \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": container with ID starting with 0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.014348    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.014663    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"} err="failed to get container status \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": rpc error: code = NotFound desc = could not find container \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": container with ID starting with f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.014698    4824 scope.go:117] "RemoveContainer" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.014972    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"} err="failed to get container status \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": rpc error: code = NotFound desc = could not find container \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": container with ID starting with 20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.015008    4824 scope.go:117] "RemoveContainer" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.015534    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"} err="failed to get container status \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": rpc error: code = NotFound desc = could not find container \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": container with ID starting with 37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.015561    4824 scope.go:117] "RemoveContainer" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.016055    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"} err="failed to get container status \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": rpc error: code = NotFound desc = could not find container \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": container with ID starting with 070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.016082    4824 scope.go:117] "RemoveContainer" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.016494    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"} err="failed to get container status \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": rpc error: code = NotFound desc = could not find container \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": container with ID starting with 0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.016650    4824 scope.go:117] "RemoveContainer" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.017196    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"} err="failed to get container status \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": rpc error: code = NotFound desc = could not find container \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": container with ID starting with 466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.017227    4824 scope.go:117] "RemoveContainer" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.017674    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"} err="failed to get container status \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": rpc error: code = NotFound desc = could not find container \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": container with ID starting with fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.017706    4824 scope.go:117] "RemoveContainer" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.018027    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"} err="failed to get container status \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": rpc error: code = NotFound desc = could not find container \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": container with ID starting with 430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.018146    4824 scope.go:117] "RemoveContainer" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.018820    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"} err="failed to get container status \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": rpc error: code = NotFound desc = could not find container \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": container with ID starting with e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.018943    4824 scope.go:117] "RemoveContainer" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.019347    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"} err="failed to get container status \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": rpc error: code = NotFound desc = could not find container \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": container with ID starting with 0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.019398    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.019857    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"} err="failed to get container status \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": rpc error: code = NotFound desc = could not find container \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": container with ID starting with f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.020083    4824 scope.go:117] "RemoveContainer" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.020678    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"} err="failed to get container status \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": rpc error: code = NotFound desc = could not find container \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": container with ID starting with 20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.020834    4824 scope.go:117] "RemoveContainer" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.021268    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"} err="failed to get container status \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": rpc error: code = NotFound desc = could not find container \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": container with ID starting with 37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.021428    4824 scope.go:117] "RemoveContainer" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.022241    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"} err="failed to get container status \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": rpc error: code = NotFound desc = could not find container \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": container with ID starting with 070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.022405    4824 scope.go:117] "RemoveContainer" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.022921    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"} err="failed to get container status \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": rpc error: code = NotFound desc = could not find container \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": container with ID starting with 0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.023078    4824 scope.go:117] "RemoveContainer" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.023579    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"} err="failed to get container status \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": rpc error: code = NotFound desc = could not find container \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": container with ID starting with 466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.023630    4824 scope.go:117] "RemoveContainer" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.023989    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"} err="failed to get container status \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": rpc error: code = NotFound desc = could not find container \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": container with ID starting with fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.024110    4824 scope.go:117] "RemoveContainer" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.024569    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"} err="failed to get container status \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": rpc error: code = NotFound desc = could not find container \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": container with ID starting with 430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.024612    4824 scope.go:117] "RemoveContainer" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.025027    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"} err="failed to get container status \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": rpc error: code = NotFound desc = could not find container \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": container with ID starting with e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.025146    4824 scope.go:117] "RemoveContainer" containerID="0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.025534    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5"} err="failed to get container status \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": rpc error: code = NotFound desc = could not find container \"0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5\": container with ID starting with 0e25828937bf32e4c48a5b01359f3b30ac64362207d8d479d8488fcd1f16fab5 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.025564    4824 scope.go:117] "RemoveContainer" containerID="f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.025798    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646"} err="failed to get container status \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": rpc error: code = NotFound desc = could not find container \"f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646\": container with ID starting with f8dd36ca3b90624bc91c78e21039110f2c533451add25add188f9afe00950646 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.025815    4824 scope.go:117] "RemoveContainer" containerID="20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.026262    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d"} err="failed to get container status \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": rpc error: code = NotFound desc = could not find container \"20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d\": container with ID starting with 20fa5218d945876cc0d9e0c4e5a14328b4774817a2048f53d6a458d625ca8a6d not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.026312    4824 scope.go:117] "RemoveContainer" containerID="37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.026689    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84"} err="failed to get container status \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": rpc error: code = NotFound desc = could not find container \"37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84\": container with ID starting with 37f8a0574e6f180cfbd86a7435d418a720969068dd192b8ed2b15b1817378b84 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.026817    4824 scope.go:117] "RemoveContainer" containerID="070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.028316    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7"} err="failed to get container status \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": rpc error: code = NotFound desc = could not find container \"070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7\": container with ID starting with 070dc8cb6d67e1c183d38feb1934ef054c625988df82e154b6f9b660cf4801b7 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.028346    4824 scope.go:117] "RemoveContainer" containerID="0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.028896    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524"} err="failed to get container status \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": rpc error: code = NotFound desc = could not find container \"0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524\": container with ID starting with 0d07257ee5c0daf74349020a8632ea6efc3fef38b10f3e38756c3d636c3b6524 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.028922    4824 scope.go:117] "RemoveContainer" containerID="466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.029187    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c"} err="failed to get container status \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": rpc error: code = NotFound desc = could not find container \"466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c\": container with ID starting with 466830b0455860884fb9fd803c412f9242d431b48e792bd0f480d4f79d66ba2c not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.029216    4824 scope.go:117] "RemoveContainer" containerID="fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.029583    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b"} err="failed to get container status \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": rpc error: code = NotFound desc = could not find container \"fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b\": container with ID starting with fea205e303129c340ede2dd6ec484aca1db87f27da1c94e5ae61ee39762a2a2b not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.029606    4824 scope.go:117] "RemoveContainer" containerID="430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.029908    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20"} err="failed to get container status \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": rpc error: code = NotFound desc = could not find container \"430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20\": container with ID starting with 430ca9b6e9755d251fee91f386373d2897d95f459125eeb75ab5ae9d7c776b20 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.029929    4824 scope.go:117] "RemoveContainer" containerID="e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.030248    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35"} err="failed to get container status \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": rpc error: code = NotFound desc = could not find container \"e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35\": container with ID starting with e5e281287a72ec783dc93ce0922187507090c0d45212e9351be16cd619c4dd35 not found: ID does not exist"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.477480    4824 generic.go:334] "Generic (PLEG): container finished" podID="59c1deb5-4a8b-4333-8a39-577a456beae1" containerID="3bad43fcf7c70fc640da950666dd03a86c0a2800f6a47501f4361c470693d8fe" exitCode=0
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.477574    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerDied","Data":"3bad43fcf7c70fc640da950666dd03a86c0a2800f6a47501f4361c470693d8fe"}
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.477632    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"126e84cb8a52a3b6bddda1dae9baeee98d8c4c97040ba81d6175302cbfba2a30"}
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.487508    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/2.log"
Jan 22 09:56:28 crc kubenswrapper[4824]: I0122 09:56:28.487842    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-zvvt5" event={"ID":"00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c","Type":"ContainerStarted","Data":"b5a0968524e600832448d265af8fb62c38b0a0dd9686d63b34e19bb6fc108744"}
Jan 22 09:56:29 crc kubenswrapper[4824]: I0122 09:56:29.412256    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40c9681a-fa96-4b9f-8ff5-2d81bbf190fb" path="/var/lib/kubelet/pods/40c9681a-fa96-4b9f-8ff5-2d81bbf190fb/volumes"
Jan 22 09:56:29 crc kubenswrapper[4824]: I0122 09:56:29.501176    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"cc240c3e3e3e1a5e6a31c6543ac449ce52ebbd2314f84b42df15c348443b153d"}
Jan 22 09:56:29 crc kubenswrapper[4824]: I0122 09:56:29.501221    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"8830320b7db3678eb0df067369a2b967221b393e7128bf9c23a5beaae1c4e1dd"}
Jan 22 09:56:29 crc kubenswrapper[4824]: I0122 09:56:29.501235    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"4f63da4eaafc5982251b01e26a6fbb4bfc5eb04e6944fd3098902c76d5cee62c"}
Jan 22 09:56:29 crc kubenswrapper[4824]: I0122 09:56:29.501245    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"f2929db1c7f76bb223baffdfc8a52b08376c82efbcd878c4f36aab37d121e237"}
Jan 22 09:56:30 crc kubenswrapper[4824]: I0122 09:56:30.333401    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:56:30 crc kubenswrapper[4824]: I0122 09:56:30.333781    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:56:30 crc kubenswrapper[4824]: I0122 09:56:30.510447    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"aaeb97febe995154cebaf2f755ef698adb6555f1d301c39c4ebfbe64879b5917"}
Jan 22 09:56:30 crc kubenswrapper[4824]: I0122 09:56:30.510497    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"75db47b273f631ba76009a5540738442b3248532b745ffb79e48bf9f3bbba02e"}
Jan 22 09:56:32 crc kubenswrapper[4824]: I0122 09:56:32.527223    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"40de26ccb9712239470b4e0690220dc7e200a2e6a4a5b73fb6e3b3e648104733"}
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.254000    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-2bbwc"]
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.255165    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.257486    4824 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2hlgh"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.257877    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.258324    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.258477    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.324819    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-crc-storage\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.324887    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-node-mnt\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.325027    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j94jd\" (UniqueName: \"kubernetes.io/projected/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-kube-api-access-j94jd\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.426744    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-node-mnt\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.427080    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-node-mnt\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.427189    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j94jd\" (UniqueName: \"kubernetes.io/projected/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-kube-api-access-j94jd\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.427305    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-crc-storage\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.428968    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-crc-storage\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.452141    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j94jd\" (UniqueName: \"kubernetes.io/projected/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-kube-api-access-j94jd\") pod \"crc-storage-crc-2bbwc\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") " pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: I0122 09:56:33.571419    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: E0122 09:56:33.604856    4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(3fac156c3d27d1e9ec3c73ca063acee381844593c212f3eddf2edb3f2db598f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"
Jan 22 09:56:33 crc kubenswrapper[4824]: E0122 09:56:33.604947    4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(3fac156c3d27d1e9ec3c73ca063acee381844593c212f3eddf2edb3f2db598f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: E0122 09:56:33.604984    4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(3fac156c3d27d1e9ec3c73ca063acee381844593c212f3eddf2edb3f2db598f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:33 crc kubenswrapper[4824]: E0122 09:56:33.605052    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-2bbwc_crc-storage(ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-2bbwc_crc-storage(ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(3fac156c3d27d1e9ec3c73ca063acee381844593c212f3eddf2edb3f2db598f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-2bbwc" podUID="ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.347594    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2bbwc"]
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.348192    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.348580    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:38 crc kubenswrapper[4824]: E0122 09:56:38.376216    4824 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(ba1caba9c3a55f9ebf3cfd42276684ed18ed5da37a628ae76f892b16c7652432): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"
Jan 22 09:56:38 crc kubenswrapper[4824]: E0122 09:56:38.376272    4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(ba1caba9c3a55f9ebf3cfd42276684ed18ed5da37a628ae76f892b16c7652432): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:38 crc kubenswrapper[4824]: E0122 09:56:38.376290    4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(ba1caba9c3a55f9ebf3cfd42276684ed18ed5da37a628ae76f892b16c7652432): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:38 crc kubenswrapper[4824]: E0122 09:56:38.376329    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-2bbwc_crc-storage(ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-2bbwc_crc-storage(ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-2bbwc_crc-storage_ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8_0(ba1caba9c3a55f9ebf3cfd42276684ed18ed5da37a628ae76f892b16c7652432): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-2bbwc" podUID="ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.567247    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" event={"ID":"59c1deb5-4a8b-4333-8a39-577a456beae1","Type":"ContainerStarted","Data":"904510b078606298ac6708cf29671b7a91aaeac533c4d3b9cfe29f4dad54ada1"}
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.567708    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.567726    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.567735    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.605254    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.618144    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4" podStartSLOduration=11.618114425 podStartE2EDuration="11.618114425s" podCreationTimestamp="2026-01-22 09:56:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:56:38.605725814 +0000 UTC m=+831.509202857" watchObservedRunningTime="2026-01-22 09:56:38.618114425 +0000 UTC m=+831.521591458"
Jan 22 09:56:38 crc kubenswrapper[4824]: I0122 09:56:38.629846    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:56:50 crc kubenswrapper[4824]: I0122 09:56:50.404128    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:50 crc kubenswrapper[4824]: I0122 09:56:50.405162    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:50 crc kubenswrapper[4824]: I0122 09:56:50.601664    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2bbwc"]
Jan 22 09:56:50 crc kubenswrapper[4824]: I0122 09:56:50.612128    4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Jan 22 09:56:50 crc kubenswrapper[4824]: I0122 09:56:50.654751    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2bbwc" event={"ID":"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8","Type":"ContainerStarted","Data":"2b996a34a86a458813ffd367301b3009e16016277728186a9c116aa8cbba4af1"}
Jan 22 09:56:52 crc kubenswrapper[4824]: I0122 09:56:52.666998    4824 generic.go:334] "Generic (PLEG): container finished" podID="ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" containerID="5fc3f66aa93b070c25dca60b6f2356b873e15e683caf95f3fff9d8aefd3a5072" exitCode=0
Jan 22 09:56:52 crc kubenswrapper[4824]: I0122 09:56:52.667200    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2bbwc" event={"ID":"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8","Type":"ContainerDied","Data":"5fc3f66aa93b070c25dca60b6f2356b873e15e683caf95f3fff9d8aefd3a5072"}
Jan 22 09:56:53 crc kubenswrapper[4824]: I0122 09:56:53.932778    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.105190    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-node-mnt\") pod \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") "
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.105261    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" (UID: "ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.105395    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j94jd\" (UniqueName: \"kubernetes.io/projected/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-kube-api-access-j94jd\") pod \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") "
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.105438    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-crc-storage\") pod \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\" (UID: \"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8\") "
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.105672    4824 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-node-mnt\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.110542    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-kube-api-access-j94jd" (OuterVolumeSpecName: "kube-api-access-j94jd") pod "ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" (UID: "ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8"). InnerVolumeSpecName "kube-api-access-j94jd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.126545    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" (UID: "ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.207063    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j94jd\" (UniqueName: \"kubernetes.io/projected/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-kube-api-access-j94jd\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.207111    4824 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8-crc-storage\") on node \"crc\" DevicePath \"\""
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.681171    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2bbwc" event={"ID":"ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8","Type":"ContainerDied","Data":"2b996a34a86a458813ffd367301b3009e16016277728186a9c116aa8cbba4af1"}
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.681214    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b996a34a86a458813ffd367301b3009e16016277728186a9c116aa8cbba4af1"
Jan 22 09:56:54 crc kubenswrapper[4824]: I0122 09:56:54.681245    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2bbwc"
Jan 22 09:56:57 crc kubenswrapper[4824]: I0122 09:56:57.517940    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dwgg4"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.332979    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.333296    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.333341    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.333900    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"897b49c4db7fcfddc6a831eea83804b055a91a5870ae0d9b09b9c435efb91273"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.333951    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://897b49c4db7fcfddc6a831eea83804b055a91a5870ae0d9b09b9c435efb91273" gracePeriod=600
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.712042    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="897b49c4db7fcfddc6a831eea83804b055a91a5870ae0d9b09b9c435efb91273" exitCode=0
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.712103    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"897b49c4db7fcfddc6a831eea83804b055a91a5870ae0d9b09b9c435efb91273"}
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.712447    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"830c35e94e49d8ce59cb2a03a98095eef4c11c2c5bd0a68e880b21c7e9b8661a"}
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.712469    4824 scope.go:117] "RemoveContainer" containerID="825ee924b14e8fa29e4a235d69b6ea74e054515fdadb5e9acc65891a65f1dced"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.973652    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"]
Jan 22 09:57:00 crc kubenswrapper[4824]: E0122 09:57:00.974194    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" containerName="storage"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.974211    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" containerName="storage"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.974296    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9d3639-9e7c-44b7-b7ea-5dfa2d0df5c8" containerName="storage"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.975048    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.976820    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc"
Jan 22 09:57:00 crc kubenswrapper[4824]: I0122 09:57:00.980625    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"]
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.095831    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.095879    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.095906    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-895h7\" (UniqueName: \"kubernetes.io/projected/95a0d0b9-aec9-4fba-89fd-ce6749ada765-kube-api-access-895h7\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.197498    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.197606    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.197663    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-895h7\" (UniqueName: \"kubernetes.io/projected/95a0d0b9-aec9-4fba-89fd-ce6749ada765-kube-api-access-895h7\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.198002    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.198320    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.222835    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-895h7\" (UniqueName: \"kubernetes.io/projected/95a0d0b9-aec9-4fba-89fd-ce6749ada765-kube-api-access-895h7\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.288532    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.497792    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"]
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.723200    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4" event={"ID":"95a0d0b9-aec9-4fba-89fd-ce6749ada765","Type":"ContainerStarted","Data":"dc82854d2f234c991dfc44b6d9b5e14e6bb83fbd40f1c61879ae755d3a436683"}
Jan 22 09:57:01 crc kubenswrapper[4824]: I0122 09:57:01.725002    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4" event={"ID":"95a0d0b9-aec9-4fba-89fd-ce6749ada765","Type":"ContainerStarted","Data":"8c4ae0f87b861049242b9ffad1925eb7360f865bfe3e1748ceb20be89b2405cb"}
Jan 22 09:57:02 crc kubenswrapper[4824]: I0122 09:57:02.732606    4824 generic.go:334] "Generic (PLEG): container finished" podID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerID="dc82854d2f234c991dfc44b6d9b5e14e6bb83fbd40f1c61879ae755d3a436683" exitCode=0
Jan 22 09:57:02 crc kubenswrapper[4824]: I0122 09:57:02.732654    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4" event={"ID":"95a0d0b9-aec9-4fba-89fd-ce6749ada765","Type":"ContainerDied","Data":"dc82854d2f234c991dfc44b6d9b5e14e6bb83fbd40f1c61879ae755d3a436683"}
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.352055    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rjhk5"]
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.353269    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.376224    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rjhk5"]
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.526855    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-utilities\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.526937    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-catalog-content\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.526965    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx9tj\" (UniqueName: \"kubernetes.io/projected/708d919b-631c-40da-939d-759798314e29-kube-api-access-zx9tj\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.628813    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-utilities\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.629219    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-catalog-content\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.629400    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx9tj\" (UniqueName: \"kubernetes.io/projected/708d919b-631c-40da-939d-759798314e29-kube-api-access-zx9tj\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.629264    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-utilities\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.629974    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-catalog-content\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.652505    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx9tj\" (UniqueName: \"kubernetes.io/projected/708d919b-631c-40da-939d-759798314e29-kube-api-access-zx9tj\") pod \"redhat-operators-rjhk5\" (UID: \"708d919b-631c-40da-939d-759798314e29\") " pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.699312    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:03 crc kubenswrapper[4824]: I0122 09:57:03.952097    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rjhk5"]
Jan 22 09:57:03 crc kubenswrapper[4824]: W0122 09:57:03.957523    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod708d919b_631c_40da_939d_759798314e29.slice/crio-eedf709498cc37ea5095db13eb6d3890f72af83049612879532aecedeb026fa6 WatchSource:0}: Error finding container eedf709498cc37ea5095db13eb6d3890f72af83049612879532aecedeb026fa6: Status 404 returned error can't find the container with id eedf709498cc37ea5095db13eb6d3890f72af83049612879532aecedeb026fa6
Jan 22 09:57:04 crc kubenswrapper[4824]: I0122 09:57:04.849727    4824 generic.go:334] "Generic (PLEG): container finished" podID="708d919b-631c-40da-939d-759798314e29" containerID="b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9" exitCode=0
Jan 22 09:57:04 crc kubenswrapper[4824]: I0122 09:57:04.849778    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerDied","Data":"b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9"}
Jan 22 09:57:04 crc kubenswrapper[4824]: I0122 09:57:04.850250    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerStarted","Data":"eedf709498cc37ea5095db13eb6d3890f72af83049612879532aecedeb026fa6"}
Jan 22 09:57:04 crc kubenswrapper[4824]: I0122 09:57:04.852454    4824 generic.go:334] "Generic (PLEG): container finished" podID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerID="f36b8cb9faf2e7291777a8a8734251b634299ed17069138da5988eb8e767f954" exitCode=0
Jan 22 09:57:04 crc kubenswrapper[4824]: I0122 09:57:04.852483    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4" event={"ID":"95a0d0b9-aec9-4fba-89fd-ce6749ada765","Type":"ContainerDied","Data":"f36b8cb9faf2e7291777a8a8734251b634299ed17069138da5988eb8e767f954"}
Jan 22 09:57:05 crc kubenswrapper[4824]: I0122 09:57:05.861327    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerStarted","Data":"ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe"}
Jan 22 09:57:05 crc kubenswrapper[4824]: I0122 09:57:05.863729    4824 generic.go:334] "Generic (PLEG): container finished" podID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerID="e29921f7a7b678662bef97dfda47bb9dc13edf105241ec4089bcb2c7851727ad" exitCode=0
Jan 22 09:57:05 crc kubenswrapper[4824]: I0122 09:57:05.863767    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4" event={"ID":"95a0d0b9-aec9-4fba-89fd-ce6749ada765","Type":"ContainerDied","Data":"e29921f7a7b678662bef97dfda47bb9dc13edf105241ec4089bcb2c7851727ad"}
Jan 22 09:57:06 crc kubenswrapper[4824]: I0122 09:57:06.871151    4824 generic.go:334] "Generic (PLEG): container finished" podID="708d919b-631c-40da-939d-759798314e29" containerID="ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe" exitCode=0
Jan 22 09:57:06 crc kubenswrapper[4824]: I0122 09:57:06.871203    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerDied","Data":"ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe"}
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.110193    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.279948    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-895h7\" (UniqueName: \"kubernetes.io/projected/95a0d0b9-aec9-4fba-89fd-ce6749ada765-kube-api-access-895h7\") pod \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") "
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.280039    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-bundle\") pod \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") "
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.280127    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-util\") pod \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\" (UID: \"95a0d0b9-aec9-4fba-89fd-ce6749ada765\") "
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.280756    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-bundle" (OuterVolumeSpecName: "bundle") pod "95a0d0b9-aec9-4fba-89fd-ce6749ada765" (UID: "95a0d0b9-aec9-4fba-89fd-ce6749ada765"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.288154    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a0d0b9-aec9-4fba-89fd-ce6749ada765-kube-api-access-895h7" (OuterVolumeSpecName: "kube-api-access-895h7") pod "95a0d0b9-aec9-4fba-89fd-ce6749ada765" (UID: "95a0d0b9-aec9-4fba-89fd-ce6749ada765"). InnerVolumeSpecName "kube-api-access-895h7". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.382188    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-895h7\" (UniqueName: \"kubernetes.io/projected/95a0d0b9-aec9-4fba-89fd-ce6749ada765-kube-api-access-895h7\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.382245    4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.751903    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xgclp"]
Jan 22 09:57:07 crc kubenswrapper[4824]: E0122 09:57:07.752729    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="pull"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.752814    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="pull"
Jan 22 09:57:07 crc kubenswrapper[4824]: E0122 09:57:07.752905    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="util"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.752973    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="util"
Jan 22 09:57:07 crc kubenswrapper[4824]: E0122 09:57:07.753048    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="extract"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.753122    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="extract"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.753291    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a0d0b9-aec9-4fba-89fd-ce6749ada765" containerName="extract"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.754220    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.774604    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgclp"]
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.833125    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-util" (OuterVolumeSpecName: "util") pod "95a0d0b9-aec9-4fba-89fd-ce6749ada765" (UID: "95a0d0b9-aec9-4fba-89fd-ce6749ada765"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.878511    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4" event={"ID":"95a0d0b9-aec9-4fba-89fd-ce6749ada765","Type":"ContainerDied","Data":"8c4ae0f87b861049242b9ffad1925eb7360f865bfe3e1748ceb20be89b2405cb"}
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.879506    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c4ae0f87b861049242b9ffad1925eb7360f865bfe3e1748ceb20be89b2405cb"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.878587    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.889632    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a59b526-9dc4-423e-a064-d7fd8bef974a-catalog-content\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.889684    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbcc\" (UniqueName: \"kubernetes.io/projected/6a59b526-9dc4-423e-a064-d7fd8bef974a-kube-api-access-jsbcc\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.889728    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a59b526-9dc4-423e-a064-d7fd8bef974a-utilities\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.889831    4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/95a0d0b9-aec9-4fba-89fd-ce6749ada765-util\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.990394    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a59b526-9dc4-423e-a064-d7fd8bef974a-catalog-content\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.990452    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbcc\" (UniqueName: \"kubernetes.io/projected/6a59b526-9dc4-423e-a064-d7fd8bef974a-kube-api-access-jsbcc\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.990494    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a59b526-9dc4-423e-a064-d7fd8bef974a-utilities\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.990885    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a59b526-9dc4-423e-a064-d7fd8bef974a-catalog-content\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:07 crc kubenswrapper[4824]: I0122 09:57:07.990935    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a59b526-9dc4-423e-a064-d7fd8bef974a-utilities\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.017521    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbcc\" (UniqueName: \"kubernetes.io/projected/6a59b526-9dc4-423e-a064-d7fd8bef974a-kube-api-access-jsbcc\") pod \"certified-operators-xgclp\" (UID: \"6a59b526-9dc4-423e-a064-d7fd8bef974a\") " pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.088651    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.347625    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgclp"]
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.886128    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerStarted","Data":"927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613"}
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.888024    4824 generic.go:334] "Generic (PLEG): container finished" podID="6a59b526-9dc4-423e-a064-d7fd8bef974a" containerID="831ba7831d1d8336a60fccc874dd3f123fbb3f3ce0f21ea7e225f1e99b938306" exitCode=0
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.888064    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgclp" event={"ID":"6a59b526-9dc4-423e-a064-d7fd8bef974a","Type":"ContainerDied","Data":"831ba7831d1d8336a60fccc874dd3f123fbb3f3ce0f21ea7e225f1e99b938306"}
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.888090    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgclp" event={"ID":"6a59b526-9dc4-423e-a064-d7fd8bef974a","Type":"ContainerStarted","Data":"2771e84d3fbf31fbc146690ae5a2c2e56bc575a5b370fa627fc8b3e45245dfb4"}
Jan 22 09:57:08 crc kubenswrapper[4824]: I0122 09:57:08.925316    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rjhk5" podStartSLOduration=2.770759836 podStartE2EDuration="5.925299423s" podCreationTimestamp="2026-01-22 09:57:03 +0000 UTC" firstStartedPulling="2026-01-22 09:57:04.85153128 +0000 UTC m=+857.755008273" lastFinishedPulling="2026-01-22 09:57:08.006070877 +0000 UTC m=+860.909547860" observedRunningTime="2026-01-22 09:57:08.909881762 +0000 UTC m=+861.813358745" watchObservedRunningTime="2026-01-22 09:57:08.925299423 +0000 UTC m=+861.828776416"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.636389    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-lg7pf"]
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.637695    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.640059    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.645309    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-lg7pf"]
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.645753    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-87ptc"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.647962    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.811348    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khm4j\" (UniqueName: \"kubernetes.io/projected/9e097354-ff63-42c7-ae81-c03e10980a87-kube-api-access-khm4j\") pod \"nmstate-operator-646758c888-lg7pf\" (UID: \"9e097354-ff63-42c7-ae81-c03e10980a87\") " pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.912657    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khm4j\" (UniqueName: \"kubernetes.io/projected/9e097354-ff63-42c7-ae81-c03e10980a87-kube-api-access-khm4j\") pod \"nmstate-operator-646758c888-lg7pf\" (UID: \"9e097354-ff63-42c7-ae81-c03e10980a87\") " pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.931608    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khm4j\" (UniqueName: \"kubernetes.io/projected/9e097354-ff63-42c7-ae81-c03e10980a87-kube-api-access-khm4j\") pod \"nmstate-operator-646758c888-lg7pf\" (UID: \"9e097354-ff63-42c7-ae81-c03e10980a87\") " pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf"
Jan 22 09:57:09 crc kubenswrapper[4824]: I0122 09:57:09.949564    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf"
Jan 22 09:57:10 crc kubenswrapper[4824]: I0122 09:57:10.417119    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-lg7pf"]
Jan 22 09:57:10 crc kubenswrapper[4824]: W0122 09:57:10.424975    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e097354_ff63_42c7_ae81_c03e10980a87.slice/crio-8569a997e1c8594e35dae18a596092ba8896a95aad6a75d48745058ef0162c00 WatchSource:0}: Error finding container 8569a997e1c8594e35dae18a596092ba8896a95aad6a75d48745058ef0162c00: Status 404 returned error can't find the container with id 8569a997e1c8594e35dae18a596092ba8896a95aad6a75d48745058ef0162c00
Jan 22 09:57:10 crc kubenswrapper[4824]: I0122 09:57:10.898732    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf" event={"ID":"9e097354-ff63-42c7-ae81-c03e10980a87","Type":"ContainerStarted","Data":"8569a997e1c8594e35dae18a596092ba8896a95aad6a75d48745058ef0162c00"}
Jan 22 09:57:13 crc kubenswrapper[4824]: I0122 09:57:13.700115    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:13 crc kubenswrapper[4824]: I0122 09:57:13.700497    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:14 crc kubenswrapper[4824]: I0122 09:57:14.835196    4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rjhk5" podUID="708d919b-631c-40da-939d-759798314e29" containerName="registry-server" probeResult="failure" output=<
Jan 22 09:57:14 crc kubenswrapper[4824]:         timeout: failed to connect service ":50051" within 1s
Jan 22 09:57:14 crc kubenswrapper[4824]:  >
Jan 22 09:57:20 crc kubenswrapper[4824]: I0122 09:57:20.965337    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgclp" event={"ID":"6a59b526-9dc4-423e-a064-d7fd8bef974a","Type":"ContainerStarted","Data":"d36f97d1af8440761b1e669dd5ea0aaf2e2cbbd144691b432aeacdcf88dc5cc5"}
Jan 22 09:57:20 crc kubenswrapper[4824]: I0122 09:57:20.970225    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf" event={"ID":"9e097354-ff63-42c7-ae81-c03e10980a87","Type":"ContainerStarted","Data":"afe403a053d92a91d2cfd3e56984c4b15513d02a98e1ec5ad1cd040bd4625b59"}
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.247926    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-lg7pf" podStartSLOduration=2.130472471 podStartE2EDuration="12.247905803s" podCreationTimestamp="2026-01-22 09:57:09 +0000 UTC" firstStartedPulling="2026-01-22 09:57:10.427274942 +0000 UTC m=+863.330751935" lastFinishedPulling="2026-01-22 09:57:20.544708274 +0000 UTC m=+873.448185267" observedRunningTime="2026-01-22 09:57:20.996648912 +0000 UTC m=+873.900125905" watchObservedRunningTime="2026-01-22 09:57:21.247905803 +0000 UTC m=+874.151382796"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.251570    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9bzd6"]
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.252797    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.272649    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9bzd6"]
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.347625    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-catalog-content\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.347699    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h22g\" (UniqueName: \"kubernetes.io/projected/f3307d40-ea18-49e7-ac1b-d522b5175f6a-kube-api-access-6h22g\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.347788    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-utilities\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.450374    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-catalog-content\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.451006    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-catalog-content\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.451084    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h22g\" (UniqueName: \"kubernetes.io/projected/f3307d40-ea18-49e7-ac1b-d522b5175f6a-kube-api-access-6h22g\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.451489    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-utilities\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.451860    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-utilities\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.514329    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h22g\" (UniqueName: \"kubernetes.io/projected/f3307d40-ea18-49e7-ac1b-d522b5175f6a-kube-api-access-6h22g\") pod \"community-operators-9bzd6\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") " pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:21 crc kubenswrapper[4824]: I0122 09:57:21.569212    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.127507    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.128842    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.131088    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-jkd9m"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.163282    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.186063    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2h9g\" (UniqueName: \"kubernetes.io/projected/9c6e17a7-98cc-4f0f-a999-152a81397e40-kube-api-access-k2h9g\") pod \"nmstate-metrics-54757c584b-x2bnr\" (UID: \"9c6e17a7-98cc-4f0f-a999-152a81397e40\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.221591    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.222236    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.224745    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.246411    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.250814    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-2snfz"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.251725    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.286850    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2h9g\" (UniqueName: \"kubernetes.io/projected/9c6e17a7-98cc-4f0f-a999-152a81397e40-kube-api-access-k2h9g\") pod \"nmstate-metrics-54757c584b-x2bnr\" (UID: \"9c6e17a7-98cc-4f0f-a999-152a81397e40\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.305304    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2h9g\" (UniqueName: \"kubernetes.io/projected/9c6e17a7-98cc-4f0f-a999-152a81397e40-kube-api-access-k2h9g\") pod \"nmstate-metrics-54757c584b-x2bnr\" (UID: \"9c6e17a7-98cc-4f0f-a999-152a81397e40\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.380963    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.387795    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.391649    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xpj5\" (UniqueName: \"kubernetes.io/projected/785191ef-4137-4f5b-808e-640b23400073-kube-api-access-4xpj5\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.391722    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-nmstate-lock\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.391749    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-ovs-socket\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.391802    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-dbus-socket\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.391855    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpwj2\" (UniqueName: \"kubernetes.io/projected/bec9eb1a-8712-4bd6-9e79-684b68436ef8-kube-api-access-vpwj2\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.392466    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.395056    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.395323    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.399906    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-xflrh"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.402690    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"]
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.488118    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492649    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64lq\" (UniqueName: \"kubernetes.io/projected/7a515d93-7907-4474-b22c-cce2da9f6c79-kube-api-access-q64lq\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492711    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492740    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492762    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xpj5\" (UniqueName: \"kubernetes.io/projected/785191ef-4137-4f5b-808e-640b23400073-kube-api-access-4xpj5\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492784    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-nmstate-lock\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492799    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-ovs-socket\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492822    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-dbus-socket\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492842    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7a515d93-7907-4474-b22c-cce2da9f6c79-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.492863    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpwj2\" (UniqueName: \"kubernetes.io/projected/bec9eb1a-8712-4bd6-9e79-684b68436ef8-kube-api-access-vpwj2\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:22 crc kubenswrapper[4824]: E0122 09:57:22.493175    4824 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found
Jan 22 09:57:22 crc kubenswrapper[4824]: E0122 09:57:22.493220    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair podName:bec9eb1a-8712-4bd6-9e79-684b68436ef8 nodeName:}" failed. No retries permitted until 2026-01-22 09:57:22.99320492 +0000 UTC m=+875.896681903 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-jqhws" (UID: "bec9eb1a-8712-4bd6-9e79-684b68436ef8") : secret "openshift-nmstate-webhook" not found
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.493439    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-nmstate-lock\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.493473    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-ovs-socket\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.493687    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/785191ef-4137-4f5b-808e-640b23400073-dbus-socket\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.529813    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xpj5\" (UniqueName: \"kubernetes.io/projected/785191ef-4137-4f5b-808e-640b23400073-kube-api-access-4xpj5\") pod \"nmstate-handler-2snfz\" (UID: \"785191ef-4137-4f5b-808e-640b23400073\") " pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.538485    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpwj2\" (UniqueName: \"kubernetes.io/projected/bec9eb1a-8712-4bd6-9e79-684b68436ef8-kube-api-access-vpwj2\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.570604    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.597629    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7a515d93-7907-4474-b22c-cce2da9f6c79-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.597697    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q64lq\" (UniqueName: \"kubernetes.io/projected/7a515d93-7907-4474-b22c-cce2da9f6c79-kube-api-access-q64lq\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.597766    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: E0122 09:57:22.597912    4824 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found
Jan 22 09:57:22 crc kubenswrapper[4824]: E0122 09:57:22.597966    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert podName:7a515d93-7907-4474-b22c-cce2da9f6c79 nodeName:}" failed. No retries permitted until 2026-01-22 09:57:23.097949937 +0000 UTC m=+876.001426930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-w7zbs" (UID: "7a515d93-7907-4474-b22c-cce2da9f6c79") : secret "plugin-serving-cert" not found
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.598726    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7a515d93-7907-4474-b22c-cce2da9f6c79-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.632753    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64lq\" (UniqueName: \"kubernetes.io/projected/7a515d93-7907-4474-b22c-cce2da9f6c79-kube-api-access-q64lq\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:22 crc kubenswrapper[4824]: I0122 09:57:22.677377    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9bzd6"]
Jan 22 09:57:22 crc kubenswrapper[4824]: W0122 09:57:22.796244    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3307d40_ea18_49e7_ac1b_d522b5175f6a.slice/crio-217060faf9f33665acfe2b98a6e8c9d8598243b5fd640db822648fb7a91bca48 WatchSource:0}: Error finding container 217060faf9f33665acfe2b98a6e8c9d8598243b5fd640db822648fb7a91bca48: Status 404 returned error can't find the container with id 217060faf9f33665acfe2b98a6e8c9d8598243b5fd640db822648fb7a91bca48
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.016109    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:23 crc kubenswrapper[4824]: E0122 09:57:23.016785    4824 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found
Jan 22 09:57:23 crc kubenswrapper[4824]: E0122 09:57:23.016867    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair podName:bec9eb1a-8712-4bd6-9e79-684b68436ef8 nodeName:}" failed. No retries permitted until 2026-01-22 09:57:24.016837632 +0000 UTC m=+876.920314625 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-jqhws" (UID: "bec9eb1a-8712-4bd6-9e79-684b68436ef8") : secret "openshift-nmstate-webhook" not found
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.098991    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2snfz" event={"ID":"785191ef-4137-4f5b-808e-640b23400073","Type":"ContainerStarted","Data":"3a769408257498264f9ebcad66d9a9be3c1a6666414ef6f82373374f724b0951"}
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.100109    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerStarted","Data":"217060faf9f33665acfe2b98a6e8c9d8598243b5fd640db822648fb7a91bca48"}
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.105019    4824 generic.go:334] "Generic (PLEG): container finished" podID="6a59b526-9dc4-423e-a064-d7fd8bef974a" containerID="d36f97d1af8440761b1e669dd5ea0aaf2e2cbbd144691b432aeacdcf88dc5cc5" exitCode=0
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.105050    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgclp" event={"ID":"6a59b526-9dc4-423e-a064-d7fd8bef974a","Type":"ContainerDied","Data":"d36f97d1af8440761b1e669dd5ea0aaf2e2cbbd144691b432aeacdcf88dc5cc5"}
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.120973    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:23 crc kubenswrapper[4824]: E0122 09:57:23.121118    4824 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found
Jan 22 09:57:23 crc kubenswrapper[4824]: E0122 09:57:23.121167    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert podName:7a515d93-7907-4474-b22c-cce2da9f6c79 nodeName:}" failed. No retries permitted until 2026-01-22 09:57:24.121152167 +0000 UTC m=+877.024629160 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-w7zbs" (UID: "7a515d93-7907-4474-b22c-cce2da9f6c79") : secret "plugin-serving-cert" not found
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.337564    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6f4fb9b494-n4v97"]
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.338607    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.396179    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f4fb9b494-n4v97"]
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428077    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-oauth-serving-cert\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428384    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-service-ca\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428508    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-console-config\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428603    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-trusted-ca-bundle\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428684    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbx9m\" (UniqueName: \"kubernetes.io/projected/dabf3e01-df01-448c-95cc-e182f6a703d7-kube-api-access-fbx9m\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428793    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dabf3e01-df01-448c-95cc-e182f6a703d7-console-serving-cert\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.428883    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dabf3e01-df01-448c-95cc-e182f6a703d7-console-oauth-config\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.522255    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-x2bnr"]
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530329    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbx9m\" (UniqueName: \"kubernetes.io/projected/dabf3e01-df01-448c-95cc-e182f6a703d7-kube-api-access-fbx9m\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530410    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dabf3e01-df01-448c-95cc-e182f6a703d7-console-serving-cert\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530451    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dabf3e01-df01-448c-95cc-e182f6a703d7-console-oauth-config\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530503    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-oauth-serving-cert\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530538    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-service-ca\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530590    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-console-config\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.530615    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-trusted-ca-bundle\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.531937    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-trusted-ca-bundle\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.532456    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-service-ca\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.532698    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-console-config\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.533585    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dabf3e01-df01-448c-95cc-e182f6a703d7-oauth-serving-cert\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.537777    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dabf3e01-df01-448c-95cc-e182f6a703d7-console-oauth-config\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.538208    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dabf3e01-df01-448c-95cc-e182f6a703d7-console-serving-cert\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.552553    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbx9m\" (UniqueName: \"kubernetes.io/projected/dabf3e01-df01-448c-95cc-e182f6a703d7-kube-api-access-fbx9m\") pod \"console-6f4fb9b494-n4v97\" (UID: \"dabf3e01-df01-448c-95cc-e182f6a703d7\") " pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.692630    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.759796    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.814012    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:23 crc kubenswrapper[4824]: I0122 09:57:23.899370    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f4fb9b494-n4v97"]
Jan 22 09:57:23 crc kubenswrapper[4824]: W0122 09:57:23.979895    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddabf3e01_df01_448c_95cc_e182f6a703d7.slice/crio-ee0a0722a0bdcfb69f10219498b814a2c1b4cafef691951cca4a3a5dabfe6c77 WatchSource:0}: Error finding container ee0a0722a0bdcfb69f10219498b814a2c1b4cafef691951cca4a3a5dabfe6c77: Status 404 returned error can't find the container with id ee0a0722a0bdcfb69f10219498b814a2c1b4cafef691951cca4a3a5dabfe6c77
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.036858    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.042486    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bec9eb1a-8712-4bd6-9e79-684b68436ef8-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-jqhws\" (UID: \"bec9eb1a-8712-4bd6-9e79-684b68436ef8\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.047138    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.111529    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f4fb9b494-n4v97" event={"ID":"dabf3e01-df01-448c-95cc-e182f6a703d7","Type":"ContainerStarted","Data":"ee0a0722a0bdcfb69f10219498b814a2c1b4cafef691951cca4a3a5dabfe6c77"}
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.112600    4824 generic.go:334] "Generic (PLEG): container finished" podID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerID="b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20" exitCode=0
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.112679    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerDied","Data":"b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20"}
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.115491    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr" event={"ID":"9c6e17a7-98cc-4f0f-a999-152a81397e40","Type":"ContainerStarted","Data":"b89ecd24c61acb1b0acf11e363ba2df0d849d70949623ed9187f13f2f5a6bd0d"}
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.139400    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.143148    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a515d93-7907-4474-b22c-cce2da9f6c79-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-w7zbs\" (UID: \"7a515d93-7907-4474-b22c-cce2da9f6c79\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.214126    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.517427    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"]
Jan 22 09:57:24 crc kubenswrapper[4824]: W0122 09:57:24.532254    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbec9eb1a_8712_4bd6_9e79_684b68436ef8.slice/crio-fd2db192872e1fe685970b9f5b724f63ec27ac673ff9b777f7674b6fc497467b WatchSource:0}: Error finding container fd2db192872e1fe685970b9f5b724f63ec27ac673ff9b777f7674b6fc497467b: Status 404 returned error can't find the container with id fd2db192872e1fe685970b9f5b724f63ec27ac673ff9b777f7674b6fc497467b
Jan 22 09:57:24 crc kubenswrapper[4824]: I0122 09:57:24.638677    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs"]
Jan 22 09:57:25 crc kubenswrapper[4824]: I0122 09:57:25.187787    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws" event={"ID":"bec9eb1a-8712-4bd6-9e79-684b68436ef8","Type":"ContainerStarted","Data":"fd2db192872e1fe685970b9f5b724f63ec27ac673ff9b777f7674b6fc497467b"}
Jan 22 09:57:25 crc kubenswrapper[4824]: I0122 09:57:25.190573    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f4fb9b494-n4v97" event={"ID":"dabf3e01-df01-448c-95cc-e182f6a703d7","Type":"ContainerStarted","Data":"09933442125ae4e9b9659dae6baa6ce24167f75325d4f0c95c3736bf4dcb9b99"}
Jan 22 09:57:25 crc kubenswrapper[4824]: I0122 09:57:25.191901    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs" event={"ID":"7a515d93-7907-4474-b22c-cce2da9f6c79","Type":"ContainerStarted","Data":"86238f26a83dbd15d3e23a134fa8e6854fccb14fe562f843d1cdb702b4b57130"}
Jan 22 09:57:25 crc kubenswrapper[4824]: I0122 09:57:25.195185    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgclp" event={"ID":"6a59b526-9dc4-423e-a064-d7fd8bef974a","Type":"ContainerStarted","Data":"1263990366aabb6c11822e1b135fd0cd0d9f576abeecdc5ce62dbf85657ab4ca"}
Jan 22 09:57:25 crc kubenswrapper[4824]: I0122 09:57:25.231602    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6f4fb9b494-n4v97" podStartSLOduration=2.231579228 podStartE2EDuration="2.231579228s" podCreationTimestamp="2026-01-22 09:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:57:25.228348768 +0000 UTC m=+878.131825781" watchObservedRunningTime="2026-01-22 09:57:25.231579228 +0000 UTC m=+878.135056221"
Jan 22 09:57:25 crc kubenswrapper[4824]: I0122 09:57:25.252138    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xgclp" podStartSLOduration=3.260459081 podStartE2EDuration="18.252111002s" podCreationTimestamp="2026-01-22 09:57:07 +0000 UTC" firstStartedPulling="2026-01-22 09:57:08.889952695 +0000 UTC m=+861.793429688" lastFinishedPulling="2026-01-22 09:57:23.881604616 +0000 UTC m=+876.785081609" observedRunningTime="2026-01-22 09:57:25.245220269 +0000 UTC m=+878.148697262" watchObservedRunningTime="2026-01-22 09:57:25.252111002 +0000 UTC m=+878.155588015"
Jan 22 09:57:27 crc kubenswrapper[4824]: I0122 09:57:27.396612    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rjhk5"]
Jan 22 09:57:27 crc kubenswrapper[4824]: I0122 09:57:27.397259    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rjhk5" podUID="708d919b-631c-40da-939d-759798314e29" containerName="registry-server" containerID="cri-o://927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613" gracePeriod=2
Jan 22 09:57:28 crc kubenswrapper[4824]: I0122 09:57:28.089725    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:28 crc kubenswrapper[4824]: I0122 09:57:28.089832    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:28 crc kubenswrapper[4824]: I0122 09:57:28.145882    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.233327    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.262271    4824 generic.go:334] "Generic (PLEG): container finished" podID="708d919b-631c-40da-939d-759798314e29" containerID="927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613" exitCode=0
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.262325    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerDied","Data":"927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613"}
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.262398    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rjhk5" event={"ID":"708d919b-631c-40da-939d-759798314e29","Type":"ContainerDied","Data":"eedf709498cc37ea5095db13eb6d3890f72af83049612879532aecedeb026fa6"}
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.262419    4824 scope.go:117] "RemoveContainer" containerID="927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.262351    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rjhk5"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.279075    4824 scope.go:117] "RemoveContainer" containerID="ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.293991    4824 scope.go:117] "RemoveContainer" containerID="b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.313917    4824 scope.go:117] "RemoveContainer" containerID="927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613"
Jan 22 09:57:29 crc kubenswrapper[4824]: E0122 09:57:29.314325    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613\": container with ID starting with 927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613 not found: ID does not exist" containerID="927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.314357    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613"} err="failed to get container status \"927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613\": rpc error: code = NotFound desc = could not find container \"927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613\": container with ID starting with 927e9a0a038fa93c854db4fe681d6c163ed8f7546edaeb92a3507aac06615613 not found: ID does not exist"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.314398    4824 scope.go:117] "RemoveContainer" containerID="ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe"
Jan 22 09:57:29 crc kubenswrapper[4824]: E0122 09:57:29.314943    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe\": container with ID starting with ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe not found: ID does not exist" containerID="ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.314986    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe"} err="failed to get container status \"ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe\": rpc error: code = NotFound desc = could not find container \"ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe\": container with ID starting with ba72ad31839a537d20277e447caabc89dfc8df84461c081425de5a9fd001fafe not found: ID does not exist"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.315017    4824 scope.go:117] "RemoveContainer" containerID="b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9"
Jan 22 09:57:29 crc kubenswrapper[4824]: E0122 09:57:29.315568    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9\": container with ID starting with b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9 not found: ID does not exist" containerID="b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.315589    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9"} err="failed to get container status \"b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9\": rpc error: code = NotFound desc = could not find container \"b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9\": container with ID starting with b21fec838f0d2e1157f6d233786f3d132d17646ee0f5a1ddfd77200b5046b3d9 not found: ID does not exist"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.345327    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-catalog-content\") pod \"708d919b-631c-40da-939d-759798314e29\" (UID: \"708d919b-631c-40da-939d-759798314e29\") "
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.345445    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx9tj\" (UniqueName: \"kubernetes.io/projected/708d919b-631c-40da-939d-759798314e29-kube-api-access-zx9tj\") pod \"708d919b-631c-40da-939d-759798314e29\" (UID: \"708d919b-631c-40da-939d-759798314e29\") "
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.345478    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-utilities\") pod \"708d919b-631c-40da-939d-759798314e29\" (UID: \"708d919b-631c-40da-939d-759798314e29\") "
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.346780    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-utilities" (OuterVolumeSpecName: "utilities") pod "708d919b-631c-40da-939d-759798314e29" (UID: "708d919b-631c-40da-939d-759798314e29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.350991    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708d919b-631c-40da-939d-759798314e29-kube-api-access-zx9tj" (OuterVolumeSpecName: "kube-api-access-zx9tj") pod "708d919b-631c-40da-939d-759798314e29" (UID: "708d919b-631c-40da-939d-759798314e29"). InnerVolumeSpecName "kube-api-access-zx9tj". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.446890    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx9tj\" (UniqueName: \"kubernetes.io/projected/708d919b-631c-40da-939d-759798314e29-kube-api-access-zx9tj\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.446923    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.468860    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "708d919b-631c-40da-939d-759798314e29" (UID: "708d919b-631c-40da-939d-759798314e29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.548094    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/708d919b-631c-40da-939d-759798314e29-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.606644    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rjhk5"]
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.616016    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rjhk5"]
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.799141    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tsv6d"]
Jan 22 09:57:29 crc kubenswrapper[4824]: E0122 09:57:29.799617    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708d919b-631c-40da-939d-759798314e29" containerName="registry-server"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.799682    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="708d919b-631c-40da-939d-759798314e29" containerName="registry-server"
Jan 22 09:57:29 crc kubenswrapper[4824]: E0122 09:57:29.799745    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708d919b-631c-40da-939d-759798314e29" containerName="extract-content"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.799807    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="708d919b-631c-40da-939d-759798314e29" containerName="extract-content"
Jan 22 09:57:29 crc kubenswrapper[4824]: E0122 09:57:29.799900    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708d919b-631c-40da-939d-759798314e29" containerName="extract-utilities"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.799986    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="708d919b-631c-40da-939d-759798314e29" containerName="extract-utilities"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.800171    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="708d919b-631c-40da-939d-759798314e29" containerName="registry-server"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.801164    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.812360    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsv6d"]
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.952998    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5tv4\" (UniqueName: \"kubernetes.io/projected/00d71a61-4d66-4baf-a033-2c930926342a-kube-api-access-b5tv4\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.953051    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-catalog-content\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:29 crc kubenswrapper[4824]: I0122 09:57:29.953086    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-utilities\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.054962    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5tv4\" (UniqueName: \"kubernetes.io/projected/00d71a61-4d66-4baf-a033-2c930926342a-kube-api-access-b5tv4\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.055405    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-catalog-content\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.055454    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-utilities\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.055936    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-utilities\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.056195    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-catalog-content\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.074984    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5tv4\" (UniqueName: \"kubernetes.io/projected/00d71a61-4d66-4baf-a033-2c930926342a-kube-api-access-b5tv4\") pod \"redhat-marketplace-tsv6d\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") " pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.120153    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:30.269546    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws" event={"ID":"bec9eb1a-8712-4bd6-9e79-684b68436ef8","Type":"ContainerStarted","Data":"037e686c185bec24e84254f2761e85cb0ac9c8ddf9bc94d6a70fd6e558cb9438"}
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:31.276416    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerStarted","Data":"284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5"}
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:31.277678    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr" event={"ID":"9c6e17a7-98cc-4f0f-a999-152a81397e40","Type":"ContainerStarted","Data":"4bfbe66843c74a5156f2a7dee6625070389c2af6195f90cb905be7047d51ace0"}
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:31.410837    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708d919b-631c-40da-939d-759798314e29" path="/var/lib/kubelet/pods/708d919b-631c-40da-939d-759798314e29/volumes"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:32.286441    4824 generic.go:334] "Generic (PLEG): container finished" podID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerID="284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5" exitCode=0
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:32.286571    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerDied","Data":"284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5"}
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:32.286647    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:32.304146    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws" podStartSLOduration=5.8741372080000005 podStartE2EDuration="10.304130895s" podCreationTimestamp="2026-01-22 09:57:22 +0000 UTC" firstStartedPulling="2026-01-22 09:57:24.535868767 +0000 UTC m=+877.439345760" lastFinishedPulling="2026-01-22 09:57:28.965862454 +0000 UTC m=+881.869339447" observedRunningTime="2026-01-22 09:57:32.301392948 +0000 UTC m=+885.204869981" watchObservedRunningTime="2026-01-22 09:57:32.304130895 +0000 UTC m=+885.207607888"
Jan 22 09:57:32 crc kubenswrapper[4824]: I0122 09:57:32.963773    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsv6d"]
Jan 22 09:57:33 crc kubenswrapper[4824]: I0122 09:57:33.294321    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsv6d" event={"ID":"00d71a61-4d66-4baf-a033-2c930926342a","Type":"ContainerStarted","Data":"e69630e314170001dd19da09ec5ca29bd5f7bb0c7eea78a278e65781f433f716"}
Jan 22 09:57:33 crc kubenswrapper[4824]: I0122 09:57:33.693015    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:33 crc kubenswrapper[4824]: I0122 09:57:33.693165    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:33 crc kubenswrapper[4824]: I0122 09:57:33.698896    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:34 crc kubenswrapper[4824]: I0122 09:57:34.300423    4824 generic.go:334] "Generic (PLEG): container finished" podID="00d71a61-4d66-4baf-a033-2c930926342a" containerID="8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127" exitCode=0
Jan 22 09:57:34 crc kubenswrapper[4824]: I0122 09:57:34.301635    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsv6d" event={"ID":"00d71a61-4d66-4baf-a033-2c930926342a","Type":"ContainerDied","Data":"8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127"}
Jan 22 09:57:34 crc kubenswrapper[4824]: I0122 09:57:34.305063    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6f4fb9b494-n4v97"
Jan 22 09:57:34 crc kubenswrapper[4824]: I0122 09:57:34.394126    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8878r"]
Jan 22 09:57:35 crc kubenswrapper[4824]: I0122 09:57:35.307566    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2snfz" event={"ID":"785191ef-4137-4f5b-808e-640b23400073","Type":"ContainerStarted","Data":"70e28e874bfbf1b531ffbebbc5bdbbf6a67855be28bc4f913ded6a9d401f0f03"}
Jan 22 09:57:35 crc kubenswrapper[4824]: I0122 09:57:35.308098    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:35 crc kubenswrapper[4824]: I0122 09:57:35.311406    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerStarted","Data":"eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4"}
Jan 22 09:57:35 crc kubenswrapper[4824]: I0122 09:57:35.312889    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs" event={"ID":"7a515d93-7907-4474-b22c-cce2da9f6c79","Type":"ContainerStarted","Data":"27e498b39a1c419090f7825b47c749c7a0119fc5a4fdab12f2b442df2f86c000"}
Jan 22 09:57:35 crc kubenswrapper[4824]: I0122 09:57:35.330155    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-2snfz" podStartSLOduration=1.624866441 podStartE2EDuration="13.33013638s" podCreationTimestamp="2026-01-22 09:57:22 +0000 UTC" firstStartedPulling="2026-01-22 09:57:22.750574892 +0000 UTC m=+875.654051895" lastFinishedPulling="2026-01-22 09:57:34.455844841 +0000 UTC m=+887.359321834" observedRunningTime="2026-01-22 09:57:35.325879891 +0000 UTC m=+888.229356884" watchObservedRunningTime="2026-01-22 09:57:35.33013638 +0000 UTC m=+888.233613373"
Jan 22 09:57:36 crc kubenswrapper[4824]: I0122 09:57:36.321312    4824 generic.go:334] "Generic (PLEG): container finished" podID="00d71a61-4d66-4baf-a033-2c930926342a" containerID="9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a" exitCode=0
Jan 22 09:57:36 crc kubenswrapper[4824]: I0122 09:57:36.321470    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsv6d" event={"ID":"00d71a61-4d66-4baf-a033-2c930926342a","Type":"ContainerDied","Data":"9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a"}
Jan 22 09:57:36 crc kubenswrapper[4824]: I0122 09:57:36.353063    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-w7zbs" podStartSLOduration=3.955093195 podStartE2EDuration="14.353042883s" podCreationTimestamp="2026-01-22 09:57:22 +0000 UTC" firstStartedPulling="2026-01-22 09:57:24.660644474 +0000 UTC m=+877.564121467" lastFinishedPulling="2026-01-22 09:57:35.058594162 +0000 UTC m=+887.962071155" observedRunningTime="2026-01-22 09:57:36.350613815 +0000 UTC m=+889.254090838" watchObservedRunningTime="2026-01-22 09:57:36.353042883 +0000 UTC m=+889.256519876"
Jan 22 09:57:36 crc kubenswrapper[4824]: I0122 09:57:36.378667    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9bzd6" podStartSLOduration=4.434103678 podStartE2EDuration="15.378644849s" podCreationTimestamp="2026-01-22 09:57:21 +0000 UTC" firstStartedPulling="2026-01-22 09:57:24.114258527 +0000 UTC m=+877.017735520" lastFinishedPulling="2026-01-22 09:57:35.058799698 +0000 UTC m=+887.962276691" observedRunningTime="2026-01-22 09:57:36.375317206 +0000 UTC m=+889.278794219" watchObservedRunningTime="2026-01-22 09:57:36.378644849 +0000 UTC m=+889.282121842"
Jan 22 09:57:38 crc kubenswrapper[4824]: I0122 09:57:38.136044    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xgclp"
Jan 22 09:57:39 crc kubenswrapper[4824]: I0122 09:57:39.340379    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsv6d" event={"ID":"00d71a61-4d66-4baf-a033-2c930926342a","Type":"ContainerStarted","Data":"047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606"}
Jan 22 09:57:39 crc kubenswrapper[4824]: I0122 09:57:39.342793    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr" event={"ID":"9c6e17a7-98cc-4f0f-a999-152a81397e40","Type":"ContainerStarted","Data":"07eba79b4e4c68a1dce01d2f171fd799faeff54cd2ec923c96fc744898ad8952"}
Jan 22 09:57:39 crc kubenswrapper[4824]: I0122 09:57:39.359351    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tsv6d" podStartSLOduration=6.504616396 podStartE2EDuration="10.359329128s" podCreationTimestamp="2026-01-22 09:57:29 +0000 UTC" firstStartedPulling="2026-01-22 09:57:34.321050304 +0000 UTC m=+887.224527297" lastFinishedPulling="2026-01-22 09:57:38.175763036 +0000 UTC m=+891.079240029" observedRunningTime="2026-01-22 09:57:39.357839026 +0000 UTC m=+892.261316029" watchObservedRunningTime="2026-01-22 09:57:39.359329128 +0000 UTC m=+892.262806121"
Jan 22 09:57:39 crc kubenswrapper[4824]: I0122 09:57:39.830413    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-x2bnr" podStartSLOduration=3.179512833 podStartE2EDuration="17.830389071s" podCreationTimestamp="2026-01-22 09:57:22 +0000 UTC" firstStartedPulling="2026-01-22 09:57:23.526059751 +0000 UTC m=+876.429536744" lastFinishedPulling="2026-01-22 09:57:38.176935989 +0000 UTC m=+891.080412982" observedRunningTime="2026-01-22 09:57:39.379300546 +0000 UTC m=+892.282777539" watchObservedRunningTime="2026-01-22 09:57:39.830389071 +0000 UTC m=+892.733866094"
Jan 22 09:57:39 crc kubenswrapper[4824]: I0122 09:57:39.831747    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgclp"]
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.120818    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.120877    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.170526    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.197415    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pb5fd"]
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.197715    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pb5fd" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="registry-server" containerID="cri-o://fca4a3d5e4e5ed5496d18732969eb01f42d50954ed6aaf181cc79205888870b2" gracePeriod=2
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.349407    4824 generic.go:334] "Generic (PLEG): container finished" podID="72d1dd96-6496-4073-b2db-de3451546d7c" containerID="fca4a3d5e4e5ed5496d18732969eb01f42d50954ed6aaf181cc79205888870b2" exitCode=0
Jan 22 09:57:40 crc kubenswrapper[4824]: I0122 09:57:40.349481    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pb5fd" event={"ID":"72d1dd96-6496-4073-b2db-de3451546d7c","Type":"ContainerDied","Data":"fca4a3d5e4e5ed5496d18732969eb01f42d50954ed6aaf181cc79205888870b2"}
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.093718    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.208791    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-utilities\") pod \"72d1dd96-6496-4073-b2db-de3451546d7c\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") "
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.208897    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-catalog-content\") pod \"72d1dd96-6496-4073-b2db-de3451546d7c\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") "
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.208958    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxhqm\" (UniqueName: \"kubernetes.io/projected/72d1dd96-6496-4073-b2db-de3451546d7c-kube-api-access-gxhqm\") pod \"72d1dd96-6496-4073-b2db-de3451546d7c\" (UID: \"72d1dd96-6496-4073-b2db-de3451546d7c\") "
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.209969    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-utilities" (OuterVolumeSpecName: "utilities") pod "72d1dd96-6496-4073-b2db-de3451546d7c" (UID: "72d1dd96-6496-4073-b2db-de3451546d7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.216913    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d1dd96-6496-4073-b2db-de3451546d7c-kube-api-access-gxhqm" (OuterVolumeSpecName: "kube-api-access-gxhqm") pod "72d1dd96-6496-4073-b2db-de3451546d7c" (UID: "72d1dd96-6496-4073-b2db-de3451546d7c"). InnerVolumeSpecName "kube-api-access-gxhqm". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.261277    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72d1dd96-6496-4073-b2db-de3451546d7c" (UID: "72d1dd96-6496-4073-b2db-de3451546d7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.310319    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.310347    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxhqm\" (UniqueName: \"kubernetes.io/projected/72d1dd96-6496-4073-b2db-de3451546d7c-kube-api-access-gxhqm\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.310386    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72d1dd96-6496-4073-b2db-de3451546d7c-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.356901    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pb5fd"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.359492    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pb5fd" event={"ID":"72d1dd96-6496-4073-b2db-de3451546d7c","Type":"ContainerDied","Data":"c8386719eeeeeb0778791f9fc7b212358c0bf5abe62fb48bdd10c32fe9f29403"}
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.359542    4824 scope.go:117] "RemoveContainer" containerID="fca4a3d5e4e5ed5496d18732969eb01f42d50954ed6aaf181cc79205888870b2"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.382279    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pb5fd"]
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.385026    4824 scope.go:117] "RemoveContainer" containerID="cfb02a68f1f8b38b043e4204a819a352d1036210349ce3e0e252c02259e68c99"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.385435    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pb5fd"]
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.409985    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" path="/var/lib/kubelet/pods/72d1dd96-6496-4073-b2db-de3451546d7c/volumes"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.418606    4824 scope.go:117] "RemoveContainer" containerID="1820fa629419cbc44acc407648aab895ae74917da5e988f704c2af16b621d95d"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.569723    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.569782    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:41 crc kubenswrapper[4824]: I0122 09:57:41.609726    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:42 crc kubenswrapper[4824]: I0122 09:57:42.405887    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:42 crc kubenswrapper[4824]: I0122 09:57:42.593137    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-2snfz"
Jan 22 09:57:44 crc kubenswrapper[4824]: I0122 09:57:44.054293    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-jqhws"
Jan 22 09:57:44 crc kubenswrapper[4824]: I0122 09:57:44.594766    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9bzd6"]
Jan 22 09:57:44 crc kubenswrapper[4824]: I0122 09:57:44.594988    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9bzd6" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="registry-server" containerID="cri-o://eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4" gracePeriod=2
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.028769    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.073581    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-utilities\") pod \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") "
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.073662    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h22g\" (UniqueName: \"kubernetes.io/projected/f3307d40-ea18-49e7-ac1b-d522b5175f6a-kube-api-access-6h22g\") pod \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") "
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.073687    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-catalog-content\") pod \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\" (UID: \"f3307d40-ea18-49e7-ac1b-d522b5175f6a\") "
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.074835    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-utilities" (OuterVolumeSpecName: "utilities") pod "f3307d40-ea18-49e7-ac1b-d522b5175f6a" (UID: "f3307d40-ea18-49e7-ac1b-d522b5175f6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.081650    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3307d40-ea18-49e7-ac1b-d522b5175f6a-kube-api-access-6h22g" (OuterVolumeSpecName: "kube-api-access-6h22g") pod "f3307d40-ea18-49e7-ac1b-d522b5175f6a" (UID: "f3307d40-ea18-49e7-ac1b-d522b5175f6a"). InnerVolumeSpecName "kube-api-access-6h22g". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.120435    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3307d40-ea18-49e7-ac1b-d522b5175f6a" (UID: "f3307d40-ea18-49e7-ac1b-d522b5175f6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.175231    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.175512    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h22g\" (UniqueName: \"kubernetes.io/projected/f3307d40-ea18-49e7-ac1b-d522b5175f6a-kube-api-access-6h22g\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.175620    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3307d40-ea18-49e7-ac1b-d522b5175f6a-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.390885    4824 generic.go:334] "Generic (PLEG): container finished" podID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerID="eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4" exitCode=0
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.390926    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerDied","Data":"eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4"}
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.390952    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9bzd6" event={"ID":"f3307d40-ea18-49e7-ac1b-d522b5175f6a","Type":"ContainerDied","Data":"217060faf9f33665acfe2b98a6e8c9d8598243b5fd640db822648fb7a91bca48"}
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.390968    4824 scope.go:117] "RemoveContainer" containerID="eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.390970    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9bzd6"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.408594    4824 scope.go:117] "RemoveContainer" containerID="284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.419013    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9bzd6"]
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.422532    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9bzd6"]
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.429074    4824 scope.go:117] "RemoveContainer" containerID="b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.446105    4824 scope.go:117] "RemoveContainer" containerID="eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4"
Jan 22 09:57:46 crc kubenswrapper[4824]: E0122 09:57:46.446703    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4\": container with ID starting with eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4 not found: ID does not exist" containerID="eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.446741    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4"} err="failed to get container status \"eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4\": rpc error: code = NotFound desc = could not find container \"eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4\": container with ID starting with eed677d181dd78569938249f3252156c5d1f1e4f88308c037505cd47c69eafb4 not found: ID does not exist"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.446766    4824 scope.go:117] "RemoveContainer" containerID="284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5"
Jan 22 09:57:46 crc kubenswrapper[4824]: E0122 09:57:46.447208    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5\": container with ID starting with 284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5 not found: ID does not exist" containerID="284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.447236    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5"} err="failed to get container status \"284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5\": rpc error: code = NotFound desc = could not find container \"284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5\": container with ID starting with 284cd94d39b2d4294eec14250248bdf74a0cbc162369200b7036b2250bef0cb5 not found: ID does not exist"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.447258    4824 scope.go:117] "RemoveContainer" containerID="b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20"
Jan 22 09:57:46 crc kubenswrapper[4824]: E0122 09:57:46.447507    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20\": container with ID starting with b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20 not found: ID does not exist" containerID="b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20"
Jan 22 09:57:46 crc kubenswrapper[4824]: I0122 09:57:46.447543    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20"} err="failed to get container status \"b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20\": rpc error: code = NotFound desc = could not find container \"b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20\": container with ID starting with b53e011c0c09cf2ff2ab9a67fc42a6a8daa48438e5c88307f6f75dc35b837f20 not found: ID does not exist"
Jan 22 09:57:47 crc kubenswrapper[4824]: I0122 09:57:47.415346    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" path="/var/lib/kubelet/pods/f3307d40-ea18-49e7-ac1b-d522b5175f6a/volumes"
Jan 22 09:57:50 crc kubenswrapper[4824]: I0122 09:57:50.167621    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:50 crc kubenswrapper[4824]: I0122 09:57:50.216975    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsv6d"]
Jan 22 09:57:50 crc kubenswrapper[4824]: I0122 09:57:50.416160    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tsv6d" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="registry-server" containerID="cri-o://047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606" gracePeriod=2
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.118069    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.158499    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-catalog-content\") pod \"00d71a61-4d66-4baf-a033-2c930926342a\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") "
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.158537    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-utilities\") pod \"00d71a61-4d66-4baf-a033-2c930926342a\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") "
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.158560    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5tv4\" (UniqueName: \"kubernetes.io/projected/00d71a61-4d66-4baf-a033-2c930926342a-kube-api-access-b5tv4\") pod \"00d71a61-4d66-4baf-a033-2c930926342a\" (UID: \"00d71a61-4d66-4baf-a033-2c930926342a\") "
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.159574    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-utilities" (OuterVolumeSpecName: "utilities") pod "00d71a61-4d66-4baf-a033-2c930926342a" (UID: "00d71a61-4d66-4baf-a033-2c930926342a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.170670    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d71a61-4d66-4baf-a033-2c930926342a-kube-api-access-b5tv4" (OuterVolumeSpecName: "kube-api-access-b5tv4") pod "00d71a61-4d66-4baf-a033-2c930926342a" (UID: "00d71a61-4d66-4baf-a033-2c930926342a"). InnerVolumeSpecName "kube-api-access-b5tv4". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.179320    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00d71a61-4d66-4baf-a033-2c930926342a" (UID: "00d71a61-4d66-4baf-a033-2c930926342a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.260107    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.260139    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d71a61-4d66-4baf-a033-2c930926342a-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.260149    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5tv4\" (UniqueName: \"kubernetes.io/projected/00d71a61-4d66-4baf-a033-2c930926342a-kube-api-access-b5tv4\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.429330    4824 generic.go:334] "Generic (PLEG): container finished" podID="00d71a61-4d66-4baf-a033-2c930926342a" containerID="047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606" exitCode=0
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.429395    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsv6d" event={"ID":"00d71a61-4d66-4baf-a033-2c930926342a","Type":"ContainerDied","Data":"047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606"}
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.429429    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tsv6d" event={"ID":"00d71a61-4d66-4baf-a033-2c930926342a","Type":"ContainerDied","Data":"e69630e314170001dd19da09ec5ca29bd5f7bb0c7eea78a278e65781f433f716"}
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.429433    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tsv6d"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.429446    4824 scope.go:117] "RemoveContainer" containerID="047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.448969    4824 scope.go:117] "RemoveContainer" containerID="9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.464165    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsv6d"]
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.468010    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tsv6d"]
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.481976    4824 scope.go:117] "RemoveContainer" containerID="8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.503533    4824 scope.go:117] "RemoveContainer" containerID="047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606"
Jan 22 09:57:52 crc kubenswrapper[4824]: E0122 09:57:52.503987    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606\": container with ID starting with 047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606 not found: ID does not exist" containerID="047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.504062    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606"} err="failed to get container status \"047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606\": rpc error: code = NotFound desc = could not find container \"047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606\": container with ID starting with 047c621fcbe1daffd7462603e5657fec075ad6679a61c9784e11c97e186ce606 not found: ID does not exist"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.504100    4824 scope.go:117] "RemoveContainer" containerID="9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a"
Jan 22 09:57:52 crc kubenswrapper[4824]: E0122 09:57:52.506086    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a\": container with ID starting with 9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a not found: ID does not exist" containerID="9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.506119    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a"} err="failed to get container status \"9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a\": rpc error: code = NotFound desc = could not find container \"9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a\": container with ID starting with 9e850d6b0b2433bd386faec0e404bb1ed589ffcaaa4c10d4595b26dd23fa593a not found: ID does not exist"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.506144    4824 scope.go:117] "RemoveContainer" containerID="8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127"
Jan 22 09:57:52 crc kubenswrapper[4824]: E0122 09:57:52.506461    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127\": container with ID starting with 8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127 not found: ID does not exist" containerID="8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127"
Jan 22 09:57:52 crc kubenswrapper[4824]: I0122 09:57:52.506485    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127"} err="failed to get container status \"8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127\": rpc error: code = NotFound desc = could not find container \"8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127\": container with ID starting with 8f22df58816940341f310a4918b23a3b21b757c76542b292f0ec0dd0e13bb127 not found: ID does not exist"
Jan 22 09:57:53 crc kubenswrapper[4824]: I0122 09:57:53.416717    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d71a61-4d66-4baf-a033-2c930926342a" path="/var/lib/kubelet/pods/00d71a61-4d66-4baf-a033-2c930926342a/volumes"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.429706    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8878r" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerName="console" containerID="cri-o://b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344" gracePeriod=15
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.800556    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8878r_cfe981a6-ed1d-411f-801d-b890b544ee5a/console/0.log"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.800629    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.851455    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-service-ca\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.851580    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-config\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.851658    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-oauth-serving-cert\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.851711    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-trusted-ca-bundle\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.851730    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tfb8\" (UniqueName: \"kubernetes.io/projected/cfe981a6-ed1d-411f-801d-b890b544ee5a-kube-api-access-5tfb8\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.852344    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-service-ca" (OuterVolumeSpecName: "service-ca") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.852376    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.852434    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-oauth-config\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.852514    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-serving-cert\") pod \"cfe981a6-ed1d-411f-801d-b890b544ee5a\" (UID: \"cfe981a6-ed1d-411f-801d-b890b544ee5a\") "
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.853015    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-config" (OuterVolumeSpecName: "console-config") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.853324    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.853462    4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-service-ca\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.853502    4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.853529    4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.857276    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.857749    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.857828    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe981a6-ed1d-411f-801d-b890b544ee5a-kube-api-access-5tfb8" (OuterVolumeSpecName: "kube-api-access-5tfb8") pod "cfe981a6-ed1d-411f-801d-b890b544ee5a" (UID: "cfe981a6-ed1d-411f-801d-b890b544ee5a"). InnerVolumeSpecName "kube-api-access-5tfb8". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935162    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"]
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935679    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerName="console"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935701    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerName="console"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935730    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935742    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935757    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935765    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935778    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="extract-utilities"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935785    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="extract-utilities"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935797    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935805    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935816    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="extract-content"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935823    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="extract-content"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935832    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="extract-content"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935838    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="extract-content"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935851    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="extract-utilities"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935858    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="extract-utilities"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935872    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="extract-content"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935879    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="extract-content"
Jan 22 09:57:59 crc kubenswrapper[4824]: E0122 09:57:59.935897    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="extract-utilities"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.935904    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="extract-utilities"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.936045    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3307d40-ea18-49e7-ac1b-d522b5175f6a" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.936070    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d1dd96-6496-4073-b2db-de3451546d7c" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.936084    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d71a61-4d66-4baf-a033-2c930926342a" containerName="registry-server"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.936096    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerName="console"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.937194    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.940285    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"]
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.942251    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc"
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.954212    4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cfe981a6-ed1d-411f-801d-b890b544ee5a-oauth-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.954238    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tfb8\" (UniqueName: \"kubernetes.io/projected/cfe981a6-ed1d-411f-801d-b890b544ee5a-kube-api-access-5tfb8\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.954252    4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-oauth-config\") on node \"crc\" DevicePath \"\""
Jan 22 09:57:59 crc kubenswrapper[4824]: I0122 09:57:59.954264    4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cfe981a6-ed1d-411f-801d-b890b544ee5a-console-serving-cert\") on node \"crc\" DevicePath \"\""
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.055840    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.056066    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck2rk\" (UniqueName: \"kubernetes.io/projected/75984dfb-1883-4993-88f7-7bb43964d1e0-kube-api-access-ck2rk\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.056138    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.157132    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.157479    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck2rk\" (UniqueName: \"kubernetes.io/projected/75984dfb-1883-4993-88f7-7bb43964d1e0-kube-api-access-ck2rk\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.157857    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.157859    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.158121    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.174783    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck2rk\" (UniqueName: \"kubernetes.io/projected/75984dfb-1883-4993-88f7-7bb43964d1e0-kube-api-access-ck2rk\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.259136    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.436704    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"]
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.473072    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w" event={"ID":"75984dfb-1883-4993-88f7-7bb43964d1e0","Type":"ContainerStarted","Data":"652ea324bece5723127a37e5e654612451ed65d2bf7d3bcc028c98c111862500"}
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.481989    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8878r_cfe981a6-ed1d-411f-801d-b890b544ee5a/console/0.log"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.482032    4824 generic.go:334] "Generic (PLEG): container finished" podID="cfe981a6-ed1d-411f-801d-b890b544ee5a" containerID="b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344" exitCode=2
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.482060    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8878r" event={"ID":"cfe981a6-ed1d-411f-801d-b890b544ee5a","Type":"ContainerDied","Data":"b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344"}
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.482082    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8878r" event={"ID":"cfe981a6-ed1d-411f-801d-b890b544ee5a","Type":"ContainerDied","Data":"d3d480caa0c4757f8d4123d3c8684149691d7c5ccdbe6205c6655e4b7402478a"}
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.482097    4824 scope.go:117] "RemoveContainer" containerID="b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.482194    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8878r"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.508294    4824 scope.go:117] "RemoveContainer" containerID="b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344"
Jan 22 09:58:00 crc kubenswrapper[4824]: E0122 09:58:00.508783    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344\": container with ID starting with b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344 not found: ID does not exist" containerID="b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.508821    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344"} err="failed to get container status \"b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344\": rpc error: code = NotFound desc = could not find container \"b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344\": container with ID starting with b4ed0439fc579ffc094ffe35471f3145f78bef7af09636702b18c8da6ab7b344 not found: ID does not exist"
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.539066    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8878r"]
Jan 22 09:58:00 crc kubenswrapper[4824]: I0122 09:58:00.542285    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8878r"]
Jan 22 09:58:01 crc kubenswrapper[4824]: I0122 09:58:01.413317    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe981a6-ed1d-411f-801d-b890b544ee5a" path="/var/lib/kubelet/pods/cfe981a6-ed1d-411f-801d-b890b544ee5a/volumes"
Jan 22 09:58:01 crc kubenswrapper[4824]: I0122 09:58:01.489905    4824 generic.go:334] "Generic (PLEG): container finished" podID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerID="73801bf773555bb0d4ab877e2a780786e42657216fa8b0f65ef8ccb30d985720" exitCode=0
Jan 22 09:58:01 crc kubenswrapper[4824]: I0122 09:58:01.489991    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w" event={"ID":"75984dfb-1883-4993-88f7-7bb43964d1e0","Type":"ContainerDied","Data":"73801bf773555bb0d4ab877e2a780786e42657216fa8b0f65ef8ccb30d985720"}
Jan 22 09:58:03 crc kubenswrapper[4824]: I0122 09:58:03.505767    4824 generic.go:334] "Generic (PLEG): container finished" podID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerID="45ecac59a153fc0092d712cd5c7ed739fa986adcd367109c0ff51140a536b8d7" exitCode=0
Jan 22 09:58:03 crc kubenswrapper[4824]: I0122 09:58:03.505826    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w" event={"ID":"75984dfb-1883-4993-88f7-7bb43964d1e0","Type":"ContainerDied","Data":"45ecac59a153fc0092d712cd5c7ed739fa986adcd367109c0ff51140a536b8d7"}
Jan 22 09:58:04 crc kubenswrapper[4824]: I0122 09:58:04.515060    4824 generic.go:334] "Generic (PLEG): container finished" podID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerID="70a79b3335ebaed84211718a4503de48f90015d15b59fb8059ba06f11f10fb06" exitCode=0
Jan 22 09:58:04 crc kubenswrapper[4824]: I0122 09:58:04.515178    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w" event={"ID":"75984dfb-1883-4993-88f7-7bb43964d1e0","Type":"ContainerDied","Data":"70a79b3335ebaed84211718a4503de48f90015d15b59fb8059ba06f11f10fb06"}
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.779574    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.832508    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-bundle\") pod \"75984dfb-1883-4993-88f7-7bb43964d1e0\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") "
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.832758    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck2rk\" (UniqueName: \"kubernetes.io/projected/75984dfb-1883-4993-88f7-7bb43964d1e0-kube-api-access-ck2rk\") pod \"75984dfb-1883-4993-88f7-7bb43964d1e0\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") "
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.832839    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-util\") pod \"75984dfb-1883-4993-88f7-7bb43964d1e0\" (UID: \"75984dfb-1883-4993-88f7-7bb43964d1e0\") "
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.833834    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-bundle" (OuterVolumeSpecName: "bundle") pod "75984dfb-1883-4993-88f7-7bb43964d1e0" (UID: "75984dfb-1883-4993-88f7-7bb43964d1e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.838450    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75984dfb-1883-4993-88f7-7bb43964d1e0-kube-api-access-ck2rk" (OuterVolumeSpecName: "kube-api-access-ck2rk") pod "75984dfb-1883-4993-88f7-7bb43964d1e0" (UID: "75984dfb-1883-4993-88f7-7bb43964d1e0"). InnerVolumeSpecName "kube-api-access-ck2rk". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.935793    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck2rk\" (UniqueName: \"kubernetes.io/projected/75984dfb-1883-4993-88f7-7bb43964d1e0-kube-api-access-ck2rk\") on node \"crc\" DevicePath \"\""
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.935896    4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:58:05 crc kubenswrapper[4824]: I0122 09:58:05.955082    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-util" (OuterVolumeSpecName: "util") pod "75984dfb-1883-4993-88f7-7bb43964d1e0" (UID: "75984dfb-1883-4993-88f7-7bb43964d1e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:58:06 crc kubenswrapper[4824]: I0122 09:58:06.037443    4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75984dfb-1883-4993-88f7-7bb43964d1e0-util\") on node \"crc\" DevicePath \"\""
Jan 22 09:58:06 crc kubenswrapper[4824]: I0122 09:58:06.530947    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w" event={"ID":"75984dfb-1883-4993-88f7-7bb43964d1e0","Type":"ContainerDied","Data":"652ea324bece5723127a37e5e654612451ed65d2bf7d3bcc028c98c111862500"}
Jan 22 09:58:06 crc kubenswrapper[4824]: I0122 09:58:06.530987    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="652ea324bece5723127a37e5e654612451ed65d2bf7d3bcc028c98c111862500"
Jan 22 09:58:06 crc kubenswrapper[4824]: I0122 09:58:06.531017    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.906596    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"]
Jan 22 09:58:14 crc kubenswrapper[4824]: E0122 09:58:14.907599    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="extract"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.907616    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="extract"
Jan 22 09:58:14 crc kubenswrapper[4824]: E0122 09:58:14.907640    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="pull"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.907646    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="pull"
Jan 22 09:58:14 crc kubenswrapper[4824]: E0122 09:58:14.907676    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="util"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.907683    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="util"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.907922    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="75984dfb-1883-4993-88f7-7bb43964d1e0" containerName="extract"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.908653    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.911794    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"]
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.912069    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.912157    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.912070    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9bhqb"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.914056    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.914268    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.945490    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j87q\" (UniqueName: \"kubernetes.io/projected/f6f5fc24-b46e-491a-a54d-380666eb1eea-kube-api-access-2j87q\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.945572    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6f5fc24-b46e-491a-a54d-380666eb1eea-apiservice-cert\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:14 crc kubenswrapper[4824]: I0122 09:58:14.945610    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6f5fc24-b46e-491a-a54d-380666eb1eea-webhook-cert\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.046284    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6f5fc24-b46e-491a-a54d-380666eb1eea-apiservice-cert\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.046378    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6f5fc24-b46e-491a-a54d-380666eb1eea-webhook-cert\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.046456    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j87q\" (UniqueName: \"kubernetes.io/projected/f6f5fc24-b46e-491a-a54d-380666eb1eea-kube-api-access-2j87q\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.051319    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6f5fc24-b46e-491a-a54d-380666eb1eea-webhook-cert\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.051380    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6f5fc24-b46e-491a-a54d-380666eb1eea-apiservice-cert\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.064178    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j87q\" (UniqueName: \"kubernetes.io/projected/f6f5fc24-b46e-491a-a54d-380666eb1eea-kube-api-access-2j87q\") pod \"metallb-operator-controller-manager-5ff4cd4548-mc6b5\" (UID: \"f6f5fc24-b46e-491a-a54d-380666eb1eea\") " pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.183283    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"]
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.184177    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.187569    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.187684    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.188073    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-mxzpg"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.207429    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"]
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.228230    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.248414    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d51ba000-c226-47ca-a8a8-19d3769e23b3-apiservice-cert\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.248510    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d51ba000-c226-47ca-a8a8-19d3769e23b3-webhook-cert\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.248582    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpwg5\" (UniqueName: \"kubernetes.io/projected/d51ba000-c226-47ca-a8a8-19d3769e23b3-kube-api-access-gpwg5\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.349986    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpwg5\" (UniqueName: \"kubernetes.io/projected/d51ba000-c226-47ca-a8a8-19d3769e23b3-kube-api-access-gpwg5\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.350326    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d51ba000-c226-47ca-a8a8-19d3769e23b3-apiservice-cert\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.350567    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d51ba000-c226-47ca-a8a8-19d3769e23b3-webhook-cert\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.374016    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpwg5\" (UniqueName: \"kubernetes.io/projected/d51ba000-c226-47ca-a8a8-19d3769e23b3-kube-api-access-gpwg5\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.375040    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d51ba000-c226-47ca-a8a8-19d3769e23b3-webhook-cert\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.378057    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d51ba000-c226-47ca-a8a8-19d3769e23b3-apiservice-cert\") pod \"metallb-operator-webhook-server-557b954cc-ldhxs\" (UID: \"d51ba000-c226-47ca-a8a8-19d3769e23b3\") " pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.500910    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.501075    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"]
Jan 22 09:58:15 crc kubenswrapper[4824]: W0122 09:58:15.515533    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6f5fc24_b46e_491a_a54d_380666eb1eea.slice/crio-3f2b0fbc98a81134b817e52905ee7ffb0635e0aac0fe8a48c83ae5b66f75a216 WatchSource:0}: Error finding container 3f2b0fbc98a81134b817e52905ee7ffb0635e0aac0fe8a48c83ae5b66f75a216: Status 404 returned error can't find the container with id 3f2b0fbc98a81134b817e52905ee7ffb0635e0aac0fe8a48c83ae5b66f75a216
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.573453    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5" event={"ID":"f6f5fc24-b46e-491a-a54d-380666eb1eea","Type":"ContainerStarted","Data":"3f2b0fbc98a81134b817e52905ee7ffb0635e0aac0fe8a48c83ae5b66f75a216"}
Jan 22 09:58:15 crc kubenswrapper[4824]: I0122 09:58:15.905250    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"]
Jan 22 09:58:15 crc kubenswrapper[4824]: W0122 09:58:15.911723    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd51ba000_c226_47ca_a8a8_19d3769e23b3.slice/crio-21fea64554dd58669c4eb6a61aa56d2faf8400b1729d3d2b7592ed2d9b24ba5b WatchSource:0}: Error finding container 21fea64554dd58669c4eb6a61aa56d2faf8400b1729d3d2b7592ed2d9b24ba5b: Status 404 returned error can't find the container with id 21fea64554dd58669c4eb6a61aa56d2faf8400b1729d3d2b7592ed2d9b24ba5b
Jan 22 09:58:16 crc kubenswrapper[4824]: I0122 09:58:16.579673    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs" event={"ID":"d51ba000-c226-47ca-a8a8-19d3769e23b3","Type":"ContainerStarted","Data":"21fea64554dd58669c4eb6a61aa56d2faf8400b1729d3d2b7592ed2d9b24ba5b"}
Jan 22 09:58:20 crc kubenswrapper[4824]: I0122 09:58:20.611688    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5" event={"ID":"f6f5fc24-b46e-491a-a54d-380666eb1eea","Type":"ContainerStarted","Data":"94b17e43c7a9c47e33b5f35cfdf88eed1d96f28ec25d68c33d14a64a23671d52"}
Jan 22 09:58:20 crc kubenswrapper[4824]: I0122 09:58:20.612334    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:20 crc kubenswrapper[4824]: I0122 09:58:20.614758    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs" event={"ID":"d51ba000-c226-47ca-a8a8-19d3769e23b3","Type":"ContainerStarted","Data":"3ad01ea8326ff1bace644465b2486ccd06c058c28ab11263cac34f4db064c122"}
Jan 22 09:58:20 crc kubenswrapper[4824]: I0122 09:58:20.614900    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:20 crc kubenswrapper[4824]: I0122 09:58:20.637062    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5" podStartSLOduration=1.780624449 podStartE2EDuration="6.637044632s" podCreationTimestamp="2026-01-22 09:58:14 +0000 UTC" firstStartedPulling="2026-01-22 09:58:15.516972741 +0000 UTC m=+928.420449744" lastFinishedPulling="2026-01-22 09:58:20.373392934 +0000 UTC m=+933.276869927" observedRunningTime="2026-01-22 09:58:20.634718367 +0000 UTC m=+933.538195360" watchObservedRunningTime="2026-01-22 09:58:20.637044632 +0000 UTC m=+933.540521625"
Jan 22 09:58:20 crc kubenswrapper[4824]: I0122 09:58:20.660241    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs" podStartSLOduration=1.187645352 podStartE2EDuration="5.660222119s" podCreationTimestamp="2026-01-22 09:58:15 +0000 UTC" firstStartedPulling="2026-01-22 09:58:15.914839169 +0000 UTC m=+928.818316162" lastFinishedPulling="2026-01-22 09:58:20.387415936 +0000 UTC m=+933.290892929" observedRunningTime="2026-01-22 09:58:20.657143893 +0000 UTC m=+933.560620896" watchObservedRunningTime="2026-01-22 09:58:20.660222119 +0000 UTC m=+933.563699112"
Jan 22 09:58:35 crc kubenswrapper[4824]: I0122 09:58:35.505930    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-557b954cc-ldhxs"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.231179    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5ff4cd4548-mc6b5"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.946625    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zd6d8"]
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.949440    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.952241    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.953030    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.953372    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hm8ks"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.968014    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"]
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.968964    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.973204    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert"
Jan 22 09:58:55 crc kubenswrapper[4824]: I0122 09:58:55.974589    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"]
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.012573    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-k77cr"]
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.013649    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.024670    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-6khr2"]
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.025682    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.026690    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-snsfw"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.026842    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.026962    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.027770    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.027870    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.033996    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-6khr2"]
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.134388    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-conf\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.134966    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-metallb-excludel2\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135039    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-startup\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135082    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics-certs\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135114    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-reloader\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135141    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135171    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps9v7\" (UniqueName: \"kubernetes.io/projected/38b26cdc-802e-4435-b3fd-e472393b9c1c-kube-api-access-ps9v7\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135218    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmhvl\" (UniqueName: \"kubernetes.io/projected/c73c646e-5500-44c3-9799-80a7f7aa6a3c-kube-api-access-bmhvl\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135248    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-sockets\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135268    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp87v\" (UniqueName: \"kubernetes.io/projected/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-kube-api-access-xp87v\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135294    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5426de92-d8af-4d6a-9264-64b224018f92-cert\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135308    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c73c646e-5500-44c3-9799-80a7f7aa6a3c-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135341    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpjtj\" (UniqueName: \"kubernetes.io/projected/5426de92-d8af-4d6a-9264-64b224018f92-kube-api-access-gpjtj\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135374    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-metrics-certs\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135394    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.135412    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5426de92-d8af-4d6a-9264-64b224018f92-metrics-certs\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.236519    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-sockets\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.236575    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp87v\" (UniqueName: \"kubernetes.io/projected/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-kube-api-access-xp87v\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.236606    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c73c646e-5500-44c3-9799-80a7f7aa6a3c-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.236624    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5426de92-d8af-4d6a-9264-64b224018f92-cert\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.236662    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpjtj\" (UniqueName: \"kubernetes.io/projected/5426de92-d8af-4d6a-9264-64b224018f92-kube-api-access-gpjtj\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.236811    4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.236893    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c73c646e-5500-44c3-9799-80a7f7aa6a3c-cert podName:c73c646e-5500-44c3-9799-80a7f7aa6a3c nodeName:}" failed. No retries permitted until 2026-01-22 09:58:56.736870427 +0000 UTC m=+969.640347420 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c73c646e-5500-44c3-9799-80a7f7aa6a3c-cert") pod "frr-k8s-webhook-server-7df86c4f6c-xrhcl" (UID: "c73c646e-5500-44c3-9799-80a7f7aa6a3c") : secret "frr-k8s-webhook-server-cert" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.236943    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-sockets\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.237252    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-metrics-certs\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.237456    4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.237517    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist podName:a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad nodeName:}" failed. No retries permitted until 2026-01-22 09:58:56.737499264 +0000 UTC m=+969.640976257 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist") pod "speaker-k77cr" (UID: "a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad") : secret "metallb-memberlist" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.237286    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238187    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5426de92-d8af-4d6a-9264-64b224018f92-metrics-certs\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238209    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-conf\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238241    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-metallb-excludel2\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238282    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-startup\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238318    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics-certs\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238338    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-reloader\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238385    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238419    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps9v7\" (UniqueName: \"kubernetes.io/projected/38b26cdc-802e-4435-b3fd-e472393b9c1c-kube-api-access-ps9v7\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238438    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmhvl\" (UniqueName: \"kubernetes.io/projected/c73c646e-5500-44c3-9799-80a7f7aa6a3c-kube-api-access-bmhvl\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238618    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-conf\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.238917    4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.238956    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics-certs podName:38b26cdc-802e-4435-b3fd-e472393b9c1c nodeName:}" failed. No retries permitted until 2026-01-22 09:58:56.738946625 +0000 UTC m=+969.642423618 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics-certs") pod "frr-k8s-zd6d8" (UID: "38b26cdc-802e-4435-b3fd-e472393b9c1c") : secret "frr-k8s-certs-secret" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.238989    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.239006    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38b26cdc-802e-4435-b3fd-e472393b9c1c-reloader\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.239072    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-metallb-excludel2\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.239477    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38b26cdc-802e-4435-b3fd-e472393b9c1c-frr-startup\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.242570    4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.243337    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-metrics-certs\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.243388    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5426de92-d8af-4d6a-9264-64b224018f92-metrics-certs\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.250753    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5426de92-d8af-4d6a-9264-64b224018f92-cert\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.259916    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp87v\" (UniqueName: \"kubernetes.io/projected/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-kube-api-access-xp87v\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.260038    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpjtj\" (UniqueName: \"kubernetes.io/projected/5426de92-d8af-4d6a-9264-64b224018f92-kube-api-access-gpjtj\") pod \"controller-6968d8fdc4-6khr2\" (UID: \"5426de92-d8af-4d6a-9264-64b224018f92\") " pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.264692    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmhvl\" (UniqueName: \"kubernetes.io/projected/c73c646e-5500-44c3-9799-80a7f7aa6a3c-kube-api-access-bmhvl\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.277954    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps9v7\" (UniqueName: \"kubernetes.io/projected/38b26cdc-802e-4435-b3fd-e472393b9c1c-kube-api-access-ps9v7\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.342617    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.743700    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics-certs\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.744098    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c73c646e-5500-44c3-9799-80a7f7aa6a3c-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.744143    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.744258    4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: E0122 09:58:56.744311    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist podName:a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad nodeName:}" failed. No retries permitted until 2026-01-22 09:58:57.744294936 +0000 UTC m=+970.647771929 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist") pod "speaker-k77cr" (UID: "a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad") : secret "metallb-memberlist" not found
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.749716    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c73c646e-5500-44c3-9799-80a7f7aa6a3c-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-xrhcl\" (UID: \"c73c646e-5500-44c3-9799-80a7f7aa6a3c\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.750103    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b26cdc-802e-4435-b3fd-e472393b9c1c-metrics-certs\") pod \"frr-k8s-zd6d8\" (UID: \"38b26cdc-802e-4435-b3fd-e472393b9c1c\") " pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.759510    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-6khr2"]
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.812697    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-6khr2" event={"ID":"5426de92-d8af-4d6a-9264-64b224018f92","Type":"ContainerStarted","Data":"127b33a3d8924cfe88aff216fa0e31df2f0df00c5c7e12f5d9107410dacdcfd4"}
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.867546    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:58:56 crc kubenswrapper[4824]: I0122 09:58:56.883112    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.228565    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"]
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.761442    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.766001    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad-memberlist\") pod \"speaker-k77cr\" (UID: \"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad\") " pod="metallb-system/speaker-k77cr"
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.829564    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl" event={"ID":"c73c646e-5500-44c3-9799-80a7f7aa6a3c","Type":"ContainerStarted","Data":"aa295e86bf563d7231f90c53085f6177206a4b254b0e0da048075da524102d97"}
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.831216    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"ec6036c6d99c28fe60625616496e07a66349924767f92588318dbccd6214036a"}
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.836010    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-6khr2" event={"ID":"5426de92-d8af-4d6a-9264-64b224018f92","Type":"ContainerStarted","Data":"1663b122eec27f7b3a5c537a37e9682829d1fb6909a524be89dec45c994dfebd"}
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.836054    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-6khr2" event={"ID":"5426de92-d8af-4d6a-9264-64b224018f92","Type":"ContainerStarted","Data":"442c6f064155168378098f2bda0fcf2057ef0cec7ef90d3cba76529abe2a73c6"}
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.836374    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.836674    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-k77cr"
Jan 22 09:58:57 crc kubenswrapper[4824]: I0122 09:58:57.873241    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-6khr2" podStartSLOduration=1.8732218409999999 podStartE2EDuration="1.873221841s" podCreationTimestamp="2026-01-22 09:58:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:58:57.869652801 +0000 UTC m=+970.773129804" watchObservedRunningTime="2026-01-22 09:58:57.873221841 +0000 UTC m=+970.776698834"
Jan 22 09:58:58 crc kubenswrapper[4824]: I0122 09:58:58.846869    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-k77cr" event={"ID":"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad","Type":"ContainerStarted","Data":"fbd7511a88920f71defe97f7cad950222595bbffa80f23e0f5664bd96dd3d655"}
Jan 22 09:58:58 crc kubenswrapper[4824]: I0122 09:58:58.847276    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-k77cr" event={"ID":"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad","Type":"ContainerStarted","Data":"3142c15d827d9fe9cc300dcfdb8604a6356f720a83f73a4667336d6b76576316"}
Jan 22 09:58:58 crc kubenswrapper[4824]: I0122 09:58:58.847293    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-k77cr" event={"ID":"a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad","Type":"ContainerStarted","Data":"f4be508e230259b239e0308aceecea945943ea12959233edf34034d97ef15066"}
Jan 22 09:58:58 crc kubenswrapper[4824]: I0122 09:58:58.869095    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-k77cr" podStartSLOduration=3.869075928 podStartE2EDuration="3.869075928s" podCreationTimestamp="2026-01-22 09:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 09:58:58.866026013 +0000 UTC m=+971.769503016" watchObservedRunningTime="2026-01-22 09:58:58.869075928 +0000 UTC m=+971.772552921"
Jan 22 09:59:00 crc kubenswrapper[4824]: I0122 09:59:00.333272    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:59:00 crc kubenswrapper[4824]: I0122 09:59:00.333690    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:59:06 crc kubenswrapper[4824]: I0122 09:59:06.346670    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-6khr2"
Jan 22 09:59:07 crc kubenswrapper[4824]: I0122 09:59:07.340959    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl" event={"ID":"c73c646e-5500-44c3-9799-80a7f7aa6a3c","Type":"ContainerStarted","Data":"c307bdf331d9a27295af4d07d95486e828878fe68fa155d69d57820a4f5702ee"}
Jan 22 09:59:07 crc kubenswrapper[4824]: I0122 09:59:07.341298    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:59:07 crc kubenswrapper[4824]: I0122 09:59:07.342741    4824 generic.go:334] "Generic (PLEG): container finished" podID="38b26cdc-802e-4435-b3fd-e472393b9c1c" containerID="aeed6c69e894b93e1623467b9a7826053a367e2feb3290b4af2e955d2eccbffd" exitCode=0
Jan 22 09:59:07 crc kubenswrapper[4824]: I0122 09:59:07.342773    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerDied","Data":"aeed6c69e894b93e1623467b9a7826053a367e2feb3290b4af2e955d2eccbffd"}
Jan 22 09:59:07 crc kubenswrapper[4824]: I0122 09:59:07.363807    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl" podStartSLOduration=2.4203865159999998 podStartE2EDuration="12.363788558s" podCreationTimestamp="2026-01-22 09:58:55 +0000 UTC" firstStartedPulling="2026-01-22 09:58:57.22827531 +0000 UTC m=+970.131752303" lastFinishedPulling="2026-01-22 09:59:07.171677332 +0000 UTC m=+980.075154345" observedRunningTime="2026-01-22 09:59:07.356328279 +0000 UTC m=+980.259805272" watchObservedRunningTime="2026-01-22 09:59:07.363788558 +0000 UTC m=+980.267265551"
Jan 22 09:59:07 crc kubenswrapper[4824]: I0122 09:59:07.836961    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-k77cr"
Jan 22 09:59:08 crc kubenswrapper[4824]: I0122 09:59:08.349797    4824 generic.go:334] "Generic (PLEG): container finished" podID="38b26cdc-802e-4435-b3fd-e472393b9c1c" containerID="e27f16e581e4bcab5364e92131c6d9bde60e52fae136a6e6b3fe64f1de93e975" exitCode=0
Jan 22 09:59:08 crc kubenswrapper[4824]: I0122 09:59:08.349861    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerDied","Data":"e27f16e581e4bcab5364e92131c6d9bde60e52fae136a6e6b3fe64f1de93e975"}
Jan 22 09:59:09 crc kubenswrapper[4824]: I0122 09:59:09.356997    4824 generic.go:334] "Generic (PLEG): container finished" podID="38b26cdc-802e-4435-b3fd-e472393b9c1c" containerID="6f2e863ef6d28e36786e5dab86663b8fcf2313ca73a86efed9a9db094dea7882" exitCode=0
Jan 22 09:59:09 crc kubenswrapper[4824]: I0122 09:59:09.357035    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerDied","Data":"6f2e863ef6d28e36786e5dab86663b8fcf2313ca73a86efed9a9db094dea7882"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370218    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"a01e21687be4b5dfbb2b97d5111717eedfe4f0e24fe2c1e296884d4356db483a"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370498    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"901cec0d737c100bb01153b09155ea1d60f4de15b5edfcb63b7c09f79af90f02"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370510    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"bed88cdd253a855a0e2fa323d28a4bbdfa42b21aa505d9756c78d6630f43dc1c"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370519    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"7990378f41200167d96c2e63cac9bb2eb0a3cf27399e17ffddfadc0f7045a0ae"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370536    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370549    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"6e3b4e373b7bf777427d0c67bfc09616773517f1fdd1fbd58ec1e614a4c4f913"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.370558    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zd6d8" event={"ID":"38b26cdc-802e-4435-b3fd-e472393b9c1c","Type":"ContainerStarted","Data":"9b10a60de68bc92d47ee4a0fd921e8c85cbefe79ec114ea9e6ff7d8e4774a584"}
Jan 22 09:59:10 crc kubenswrapper[4824]: I0122 09:59:10.393991    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zd6d8" podStartSLOduration=5.384089889 podStartE2EDuration="15.393962797s" podCreationTimestamp="2026-01-22 09:58:55 +0000 UTC" firstStartedPulling="2026-01-22 09:58:57.128206114 +0000 UTC m=+970.031683107" lastFinishedPulling="2026-01-22 09:59:07.138079002 +0000 UTC m=+980.041556015" observedRunningTime="2026-01-22 09:59:10.392633679 +0000 UTC m=+983.296110672" watchObservedRunningTime="2026-01-22 09:59:10.393962797 +0000 UTC m=+983.297439790"
Jan 22 09:59:11 crc kubenswrapper[4824]: I0122 09:59:11.869052    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:59:11 crc kubenswrapper[4824]: I0122 09:59:11.919869    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:59:17 crc kubenswrapper[4824]: I0122 09:59:17.839605    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-k77cr"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.154316    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"]
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.156067    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.158390    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.173857    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"]
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.260295    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.260439    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8qj9\" (UniqueName: \"kubernetes.io/projected/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-kube-api-access-r8qj9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.260497    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.361822    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8qj9\" (UniqueName: \"kubernetes.io/projected/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-kube-api-access-r8qj9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.361892    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.361951    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.362552    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.362575    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.383711    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8qj9\" (UniqueName: \"kubernetes.io/projected/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-kube-api-access-r8qj9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.476148    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:19 crc kubenswrapper[4824]: I0122 09:59:19.760405    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"]
Jan 22 09:59:19 crc kubenswrapper[4824]: W0122 09:59:19.771889    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfdfcbf5_dbab_420e_a5ee_17c606dba23a.slice/crio-76aff4f0c4d63cc448c1e859c6c4c59d3038e5e2306ed33cabadd271d52a0fcf WatchSource:0}: Error finding container 76aff4f0c4d63cc448c1e859c6c4c59d3038e5e2306ed33cabadd271d52a0fcf: Status 404 returned error can't find the container with id 76aff4f0c4d63cc448c1e859c6c4c59d3038e5e2306ed33cabadd271d52a0fcf
Jan 22 09:59:20 crc kubenswrapper[4824]: I0122 09:59:20.427526    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf" event={"ID":"dfdfcbf5-dbab-420e-a5ee-17c606dba23a","Type":"ContainerStarted","Data":"76aff4f0c4d63cc448c1e859c6c4c59d3038e5e2306ed33cabadd271d52a0fcf"}
Jan 22 09:59:21 crc kubenswrapper[4824]: I0122 09:59:21.436110    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf" event={"ID":"dfdfcbf5-dbab-420e-a5ee-17c606dba23a","Type":"ContainerStarted","Data":"769d726aa1e7c30dc54b1d0b1d318a8df0f46248bae3657a712d648159c9846c"}
Jan 22 09:59:22 crc kubenswrapper[4824]: I0122 09:59:22.444033    4824 generic.go:334] "Generic (PLEG): container finished" podID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerID="769d726aa1e7c30dc54b1d0b1d318a8df0f46248bae3657a712d648159c9846c" exitCode=0
Jan 22 09:59:22 crc kubenswrapper[4824]: I0122 09:59:22.444108    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf" event={"ID":"dfdfcbf5-dbab-420e-a5ee-17c606dba23a","Type":"ContainerDied","Data":"769d726aa1e7c30dc54b1d0b1d318a8df0f46248bae3657a712d648159c9846c"}
Jan 22 09:59:26 crc kubenswrapper[4824]: I0122 09:59:26.870620    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zd6d8"
Jan 22 09:59:26 crc kubenswrapper[4824]: I0122 09:59:26.894410    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-xrhcl"
Jan 22 09:59:27 crc kubenswrapper[4824]: I0122 09:59:27.474886    4824 generic.go:334] "Generic (PLEG): container finished" podID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerID="881e3f69f4f962cb1e5501da41961a4526382e59adf07199f03b68dd5f4ba527" exitCode=0
Jan 22 09:59:27 crc kubenswrapper[4824]: I0122 09:59:27.475035    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf" event={"ID":"dfdfcbf5-dbab-420e-a5ee-17c606dba23a","Type":"ContainerDied","Data":"881e3f69f4f962cb1e5501da41961a4526382e59adf07199f03b68dd5f4ba527"}
Jan 22 09:59:28 crc kubenswrapper[4824]: I0122 09:59:28.487884    4824 generic.go:334] "Generic (PLEG): container finished" podID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerID="9382ac6d65830a45d2455fa6583d59248076337aee7100e8aec1813bee58f827" exitCode=0
Jan 22 09:59:28 crc kubenswrapper[4824]: I0122 09:59:28.487942    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf" event={"ID":"dfdfcbf5-dbab-420e-a5ee-17c606dba23a","Type":"ContainerDied","Data":"9382ac6d65830a45d2455fa6583d59248076337aee7100e8aec1813bee58f827"}
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.718276    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.905265    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-bundle\") pod \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") "
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.905335    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-util\") pod \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") "
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.905426    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8qj9\" (UniqueName: \"kubernetes.io/projected/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-kube-api-access-r8qj9\") pod \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\" (UID: \"dfdfcbf5-dbab-420e-a5ee-17c606dba23a\") "
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.906666    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-bundle" (OuterVolumeSpecName: "bundle") pod "dfdfcbf5-dbab-420e-a5ee-17c606dba23a" (UID: "dfdfcbf5-dbab-420e-a5ee-17c606dba23a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.910922    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-kube-api-access-r8qj9" (OuterVolumeSpecName: "kube-api-access-r8qj9") pod "dfdfcbf5-dbab-420e-a5ee-17c606dba23a" (UID: "dfdfcbf5-dbab-420e-a5ee-17c606dba23a"). InnerVolumeSpecName "kube-api-access-r8qj9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 09:59:29 crc kubenswrapper[4824]: I0122 09:59:29.933949    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-util" (OuterVolumeSpecName: "util") pod "dfdfcbf5-dbab-420e-a5ee-17c606dba23a" (UID: "dfdfcbf5-dbab-420e-a5ee-17c606dba23a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.006578    4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-util\") on node \"crc\" DevicePath \"\""
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.006622    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8qj9\" (UniqueName: \"kubernetes.io/projected/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-kube-api-access-r8qj9\") on node \"crc\" DevicePath \"\""
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.006641    4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfdfcbf5-dbab-420e-a5ee-17c606dba23a-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.332606    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.332661    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.500534    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf" event={"ID":"dfdfcbf5-dbab-420e-a5ee-17c606dba23a","Type":"ContainerDied","Data":"76aff4f0c4d63cc448c1e859c6c4c59d3038e5e2306ed33cabadd271d52a0fcf"}
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.500598    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76aff4f0c4d63cc448c1e859c6c4c59d3038e5e2306ed33cabadd271d52a0fcf"
Jan 22 09:59:30 crc kubenswrapper[4824]: I0122 09:59:30.500598    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.200411    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"]
Jan 22 09:59:37 crc kubenswrapper[4824]: E0122 09:59:37.201158    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="pull"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.201175    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="pull"
Jan 22 09:59:37 crc kubenswrapper[4824]: E0122 09:59:37.201195    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="extract"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.201203    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="extract"
Jan 22 09:59:37 crc kubenswrapper[4824]: E0122 09:59:37.201215    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="util"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.201223    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="util"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.201349    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfdfcbf5-dbab-420e-a5ee-17c606dba23a" containerName="extract"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.202035    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.204045    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.204122    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.206598    4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-f76bz"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.216152    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"]
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.303403    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zptn8\" (UniqueName: \"kubernetes.io/projected/a4b1a01d-c755-4943-ac20-c7f3eabc2b70-kube-api-access-zptn8\") pod \"cert-manager-operator-controller-manager-64cf6dff88-nw7t8\" (UID: \"a4b1a01d-c755-4943-ac20-c7f3eabc2b70\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.303465    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a4b1a01d-c755-4943-ac20-c7f3eabc2b70-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-nw7t8\" (UID: \"a4b1a01d-c755-4943-ac20-c7f3eabc2b70\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.405121    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a4b1a01d-c755-4943-ac20-c7f3eabc2b70-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-nw7t8\" (UID: \"a4b1a01d-c755-4943-ac20-c7f3eabc2b70\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.405222    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zptn8\" (UniqueName: \"kubernetes.io/projected/a4b1a01d-c755-4943-ac20-c7f3eabc2b70-kube-api-access-zptn8\") pod \"cert-manager-operator-controller-manager-64cf6dff88-nw7t8\" (UID: \"a4b1a01d-c755-4943-ac20-c7f3eabc2b70\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.405935    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a4b1a01d-c755-4943-ac20-c7f3eabc2b70-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-nw7t8\" (UID: \"a4b1a01d-c755-4943-ac20-c7f3eabc2b70\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.423563    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zptn8\" (UniqueName: \"kubernetes.io/projected/a4b1a01d-c755-4943-ac20-c7f3eabc2b70-kube-api-access-zptn8\") pod \"cert-manager-operator-controller-manager-64cf6dff88-nw7t8\" (UID: \"a4b1a01d-c755-4943-ac20-c7f3eabc2b70\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.567669    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"
Jan 22 09:59:37 crc kubenswrapper[4824]: I0122 09:59:37.952007    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8"]
Jan 22 09:59:37 crc kubenswrapper[4824]: W0122 09:59:37.957750    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4b1a01d_c755_4943_ac20_c7f3eabc2b70.slice/crio-3405036be35f10f675a2f02e3daf9c973f5eb9a07536d65a5e8657ab1f056c1d WatchSource:0}: Error finding container 3405036be35f10f675a2f02e3daf9c973f5eb9a07536d65a5e8657ab1f056c1d: Status 404 returned error can't find the container with id 3405036be35f10f675a2f02e3daf9c973f5eb9a07536d65a5e8657ab1f056c1d
Jan 22 09:59:38 crc kubenswrapper[4824]: I0122 09:59:38.542463    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8" event={"ID":"a4b1a01d-c755-4943-ac20-c7f3eabc2b70","Type":"ContainerStarted","Data":"3405036be35f10f675a2f02e3daf9c973f5eb9a07536d65a5e8657ab1f056c1d"}
Jan 22 09:59:45 crc kubenswrapper[4824]: I0122 09:59:45.583805    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8" event={"ID":"a4b1a01d-c755-4943-ac20-c7f3eabc2b70","Type":"ContainerStarted","Data":"2b3e3b89e1257ca49ecd5bc6026598a90269d2b91be0427b5995e738efd72787"}
Jan 22 09:59:45 crc kubenswrapper[4824]: I0122 09:59:45.605824    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-nw7t8" podStartSLOduration=1.277622474 podStartE2EDuration="8.605804719s" podCreationTimestamp="2026-01-22 09:59:37 +0000 UTC" firstStartedPulling="2026-01-22 09:59:37.95903978 +0000 UTC m=+1010.862516773" lastFinishedPulling="2026-01-22 09:59:45.287222025 +0000 UTC m=+1018.190699018" observedRunningTime="2026-01-22 09:59:45.604897864 +0000 UTC m=+1018.508374857" watchObservedRunningTime="2026-01-22 09:59:45.605804719 +0000 UTC m=+1018.509281712"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.715991    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"]
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.717333    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.719062    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.719280    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.724003    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"]
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.733348    4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-w4k9k"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.772867    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tbts\" (UniqueName: \"kubernetes.io/projected/4573e1ac-2b18-48bc-93e0-03cf1f11c1b5-kube-api-access-7tbts\") pod \"cert-manager-webhook-f4fb5df64-mcc4x\" (UID: \"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.773109    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4573e1ac-2b18-48bc-93e0-03cf1f11c1b5-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-mcc4x\" (UID: \"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.874530    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tbts\" (UniqueName: \"kubernetes.io/projected/4573e1ac-2b18-48bc-93e0-03cf1f11c1b5-kube-api-access-7tbts\") pod \"cert-manager-webhook-f4fb5df64-mcc4x\" (UID: \"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.874803    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4573e1ac-2b18-48bc-93e0-03cf1f11c1b5-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-mcc4x\" (UID: \"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.893478    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4573e1ac-2b18-48bc-93e0-03cf1f11c1b5-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-mcc4x\" (UID: \"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.896999    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tbts\" (UniqueName: \"kubernetes.io/projected/4573e1ac-2b18-48bc-93e0-03cf1f11c1b5-kube-api-access-7tbts\") pod \"cert-manager-webhook-f4fb5df64-mcc4x\" (UID: \"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.926599    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"]
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.927483    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.934339    4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-zc7xb"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.934970    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"]
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.975587    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngq5\" (UniqueName: \"kubernetes.io/projected/07b9c7cd-b839-415b-ba6a-f3c21c827c7c-kube-api-access-cngq5\") pod \"cert-manager-cainjector-855d9ccff4-sfrxc\" (UID: \"07b9c7cd-b839-415b-ba6a-f3c21c827c7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:51 crc kubenswrapper[4824]: I0122 09:59:51.975675    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b9c7cd-b839-415b-ba6a-f3c21c827c7c-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-sfrxc\" (UID: \"07b9c7cd-b839-415b-ba6a-f3c21c827c7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.040499    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.076975    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b9c7cd-b839-415b-ba6a-f3c21c827c7c-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-sfrxc\" (UID: \"07b9c7cd-b839-415b-ba6a-f3c21c827c7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.077066    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngq5\" (UniqueName: \"kubernetes.io/projected/07b9c7cd-b839-415b-ba6a-f3c21c827c7c-kube-api-access-cngq5\") pod \"cert-manager-cainjector-855d9ccff4-sfrxc\" (UID: \"07b9c7cd-b839-415b-ba6a-f3c21c827c7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.101897    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b9c7cd-b839-415b-ba6a-f3c21c827c7c-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-sfrxc\" (UID: \"07b9c7cd-b839-415b-ba6a-f3c21c827c7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.119087    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngq5\" (UniqueName: \"kubernetes.io/projected/07b9c7cd-b839-415b-ba6a-f3c21c827c7c-kube-api-access-cngq5\") pod \"cert-manager-cainjector-855d9ccff4-sfrxc\" (UID: \"07b9c7cd-b839-415b-ba6a-f3c21c827c7c\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.258903    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.458449    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc"]
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.461108    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"]
Jan 22 09:59:52 crc kubenswrapper[4824]: W0122 09:59:52.467203    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07b9c7cd_b839_415b_ba6a_f3c21c827c7c.slice/crio-3867634765e9b2b88f357e83865eb978a4b6b499b0c92b7f252568977a1ae578 WatchSource:0}: Error finding container 3867634765e9b2b88f357e83865eb978a4b6b499b0c92b7f252568977a1ae578: Status 404 returned error can't find the container with id 3867634765e9b2b88f357e83865eb978a4b6b499b0c92b7f252568977a1ae578
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.763051    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc" event={"ID":"07b9c7cd-b839-415b-ba6a-f3c21c827c7c","Type":"ContainerStarted","Data":"3867634765e9b2b88f357e83865eb978a4b6b499b0c92b7f252568977a1ae578"}
Jan 22 09:59:52 crc kubenswrapper[4824]: I0122 09:59:52.764304    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x" event={"ID":"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5","Type":"ContainerStarted","Data":"888bdb5ba306dad68791a3a8343580196391e78ffeb0c8c7c7e5a37a02044f1a"}
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.469670    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-9pfww"]
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.470737    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.472257    4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bh4d8"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.482256    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvxlz\" (UniqueName: \"kubernetes.io/projected/1d78110c-c6e5-478f-a601-fe69e15d7a1d-kube-api-access-cvxlz\") pod \"cert-manager-86cb77c54b-9pfww\" (UID: \"1d78110c-c6e5-478f-a601-fe69e15d7a1d\") " pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.482320    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d78110c-c6e5-478f-a601-fe69e15d7a1d-bound-sa-token\") pod \"cert-manager-86cb77c54b-9pfww\" (UID: \"1d78110c-c6e5-478f-a601-fe69e15d7a1d\") " pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.482679    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-9pfww"]
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.584981    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d78110c-c6e5-478f-a601-fe69e15d7a1d-bound-sa-token\") pod \"cert-manager-86cb77c54b-9pfww\" (UID: \"1d78110c-c6e5-478f-a601-fe69e15d7a1d\") " pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.585497    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvxlz\" (UniqueName: \"kubernetes.io/projected/1d78110c-c6e5-478f-a601-fe69e15d7a1d-kube-api-access-cvxlz\") pod \"cert-manager-86cb77c54b-9pfww\" (UID: \"1d78110c-c6e5-478f-a601-fe69e15d7a1d\") " pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.606215    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d78110c-c6e5-478f-a601-fe69e15d7a1d-bound-sa-token\") pod \"cert-manager-86cb77c54b-9pfww\" (UID: \"1d78110c-c6e5-478f-a601-fe69e15d7a1d\") " pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.608461    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvxlz\" (UniqueName: \"kubernetes.io/projected/1d78110c-c6e5-478f-a601-fe69e15d7a1d-kube-api-access-cvxlz\") pod \"cert-manager-86cb77c54b-9pfww\" (UID: \"1d78110c-c6e5-478f-a601-fe69e15d7a1d\") " pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 09:59:58 crc kubenswrapper[4824]: I0122 09:59:58.787856    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-9pfww"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.139208    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-9pfww"]
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.145090    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"]
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.145894    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.148838    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.149066    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.160104    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"]
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.288237    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ded193f-7dc3-4123-8bcf-21f90aea83fe-secret-volume\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.288699    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z886m\" (UniqueName: \"kubernetes.io/projected/1ded193f-7dc3-4123-8bcf-21f90aea83fe-kube-api-access-z886m\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.288904    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ded193f-7dc3-4123-8bcf-21f90aea83fe-config-volume\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.332717    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.332964    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.333060    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.333643    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"830c35e94e49d8ce59cb2a03a98095eef4c11c2c5bd0a68e880b21c7e9b8661a"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.333815    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://830c35e94e49d8ce59cb2a03a98095eef4c11c2c5bd0a68e880b21c7e9b8661a" gracePeriod=600
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.390585    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ded193f-7dc3-4123-8bcf-21f90aea83fe-config-volume\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.390847    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ded193f-7dc3-4123-8bcf-21f90aea83fe-secret-volume\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.390939    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z886m\" (UniqueName: \"kubernetes.io/projected/1ded193f-7dc3-4123-8bcf-21f90aea83fe-kube-api-access-z886m\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.391475    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ded193f-7dc3-4123-8bcf-21f90aea83fe-config-volume\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.403345    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ded193f-7dc3-4123-8bcf-21f90aea83fe-secret-volume\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.414166    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z886m\" (UniqueName: \"kubernetes.io/projected/1ded193f-7dc3-4123-8bcf-21f90aea83fe-kube-api-access-z886m\") pod \"collect-profiles-29484600-mqwf6\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.496374    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.928922    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x" event={"ID":"4573e1ac-2b18-48bc-93e0-03cf1f11c1b5","Type":"ContainerStarted","Data":"e1208ce852c31158d37ceb2cd8563afd3885cd9711e2b51aaa12b063d5a7edb6"}
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.929893    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.930790    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-9pfww" event={"ID":"1d78110c-c6e5-478f-a601-fe69e15d7a1d","Type":"ContainerStarted","Data":"47152c4cb4bd841c588bfc484c6b8e4e889595e1b9e891910d4000e8fd317e68"}
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.930814    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-9pfww" event={"ID":"1d78110c-c6e5-478f-a601-fe69e15d7a1d","Type":"ContainerStarted","Data":"cefb1fa70f493dba0f1eaa5dd0a5e44e624b0a76d49c01b8eee3dd06bcda4ec9"}
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.933330    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="830c35e94e49d8ce59cb2a03a98095eef4c11c2c5bd0a68e880b21c7e9b8661a" exitCode=0
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.933386    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"830c35e94e49d8ce59cb2a03a98095eef4c11c2c5bd0a68e880b21c7e9b8661a"}
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.933407    4824 scope.go:117] "RemoveContainer" containerID="897b49c4db7fcfddc6a831eea83804b055a91a5870ae0d9b09b9c435efb91273"
Jan 22 10:00:00 crc kubenswrapper[4824]: I0122 10:00:00.952486    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc" event={"ID":"07b9c7cd-b839-415b-ba6a-f3c21c827c7c","Type":"ContainerStarted","Data":"ba0947922473297fe358d28b8ad704a632290dd3d4537f691d40cd71ec9d6930"}
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.255907    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x" podStartSLOduration=2.088300983 podStartE2EDuration="10.255893626s" podCreationTimestamp="2026-01-22 09:59:51 +0000 UTC" firstStartedPulling="2026-01-22 09:59:52.466094302 +0000 UTC m=+1025.369571295" lastFinishedPulling="2026-01-22 10:00:00.633686945 +0000 UTC m=+1033.537163938" observedRunningTime="2026-01-22 10:00:00.951348304 +0000 UTC m=+1033.854825297" watchObservedRunningTime="2026-01-22 10:00:01.255893626 +0000 UTC m=+1034.159370619"
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.286637    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-9pfww" podStartSLOduration=3.2866226960000002 podStartE2EDuration="3.286622696s" podCreationTimestamp="2026-01-22 09:59:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 10:00:01.25639581 +0000 UTC m=+1034.159872803" watchObservedRunningTime="2026-01-22 10:00:01.286622696 +0000 UTC m=+1034.190099689"
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.290536    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"]
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.366276    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-sfrxc" podStartSLOduration=2.244188994 podStartE2EDuration="10.366258214s" podCreationTimestamp="2026-01-22 09:59:51 +0000 UTC" firstStartedPulling="2026-01-22 09:59:52.468585451 +0000 UTC m=+1025.372062444" lastFinishedPulling="2026-01-22 10:00:00.590654681 +0000 UTC m=+1033.494131664" observedRunningTime="2026-01-22 10:00:01.356023768 +0000 UTC m=+1034.259500761" watchObservedRunningTime="2026-01-22 10:00:01.366258214 +0000 UTC m=+1034.269735207"
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.960438    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"142d82198208020ed04e19461fd1e7bda1b5d653c554ec42bffda74db5a4d02d"}
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.961895    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6" event={"ID":"1ded193f-7dc3-4123-8bcf-21f90aea83fe","Type":"ContainerStarted","Data":"3febecca5b1e41101c3d325ce9ed2d1e1185feb861921adedac2e797cd6dfd61"}
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.961991    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6" event={"ID":"1ded193f-7dc3-4123-8bcf-21f90aea83fe","Type":"ContainerStarted","Data":"0eda14ecd0370859ec445e3da5fc204c3f5fbff789d02641dd2d4c97094862d4"}
Jan 22 10:00:01 crc kubenswrapper[4824]: I0122 10:00:01.988069    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6" podStartSLOduration=1.988052632 podStartE2EDuration="1.988052632s" podCreationTimestamp="2026-01-22 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 10:00:01.985921902 +0000 UTC m=+1034.889398905" watchObservedRunningTime="2026-01-22 10:00:01.988052632 +0000 UTC m=+1034.891529625"
Jan 22 10:00:02 crc kubenswrapper[4824]: I0122 10:00:02.968998    4824 generic.go:334] "Generic (PLEG): container finished" podID="1ded193f-7dc3-4123-8bcf-21f90aea83fe" containerID="3febecca5b1e41101c3d325ce9ed2d1e1185feb861921adedac2e797cd6dfd61" exitCode=0
Jan 22 10:00:02 crc kubenswrapper[4824]: I0122 10:00:02.969250    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6" event={"ID":"1ded193f-7dc3-4123-8bcf-21f90aea83fe","Type":"ContainerDied","Data":"3febecca5b1e41101c3d325ce9ed2d1e1185feb861921adedac2e797cd6dfd61"}
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.306148    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.412034    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ded193f-7dc3-4123-8bcf-21f90aea83fe-secret-volume\") pod \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") "
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.412113    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z886m\" (UniqueName: \"kubernetes.io/projected/1ded193f-7dc3-4123-8bcf-21f90aea83fe-kube-api-access-z886m\") pod \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") "
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.412139    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ded193f-7dc3-4123-8bcf-21f90aea83fe-config-volume\") pod \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\" (UID: \"1ded193f-7dc3-4123-8bcf-21f90aea83fe\") "
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.413114    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ded193f-7dc3-4123-8bcf-21f90aea83fe-config-volume" (OuterVolumeSpecName: "config-volume") pod "1ded193f-7dc3-4123-8bcf-21f90aea83fe" (UID: "1ded193f-7dc3-4123-8bcf-21f90aea83fe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.417555    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ded193f-7dc3-4123-8bcf-21f90aea83fe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1ded193f-7dc3-4123-8bcf-21f90aea83fe" (UID: "1ded193f-7dc3-4123-8bcf-21f90aea83fe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.417600    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ded193f-7dc3-4123-8bcf-21f90aea83fe-kube-api-access-z886m" (OuterVolumeSpecName: "kube-api-access-z886m") pod "1ded193f-7dc3-4123-8bcf-21f90aea83fe" (UID: "1ded193f-7dc3-4123-8bcf-21f90aea83fe"). InnerVolumeSpecName "kube-api-access-z886m". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.513672    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z886m\" (UniqueName: \"kubernetes.io/projected/1ded193f-7dc3-4123-8bcf-21f90aea83fe-kube-api-access-z886m\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.513705    4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ded193f-7dc3-4123-8bcf-21f90aea83fe-config-volume\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.513714    4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ded193f-7dc3-4123-8bcf-21f90aea83fe-secret-volume\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.986672    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6" event={"ID":"1ded193f-7dc3-4123-8bcf-21f90aea83fe","Type":"ContainerDied","Data":"0eda14ecd0370859ec445e3da5fc204c3f5fbff789d02641dd2d4c97094862d4"}
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.986714    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eda14ecd0370859ec445e3da5fc204c3f5fbff789d02641dd2d4c97094862d4"
Jan 22 10:00:04 crc kubenswrapper[4824]: I0122 10:00:04.987153    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484600-mqwf6"
Jan 22 10:00:07 crc kubenswrapper[4824]: I0122 10:00:07.044676    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-mcc4x"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.925461    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t8gtv"]
Jan 22 10:00:09 crc kubenswrapper[4824]: E0122 10:00:09.926183    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ded193f-7dc3-4123-8bcf-21f90aea83fe" containerName="collect-profiles"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.926204    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ded193f-7dc3-4123-8bcf-21f90aea83fe" containerName="collect-profiles"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.926501    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ded193f-7dc3-4123-8bcf-21f90aea83fe" containerName="collect-profiles"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.927224    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.929991    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-nfkm4"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.930437    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.931403    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt"
Jan 22 10:00:09 crc kubenswrapper[4824]: I0122 10:00:09.985930    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvm9g\" (UniqueName: \"kubernetes.io/projected/8fd52a2a-b1af-4528-a6ba-2e369fed8b99-kube-api-access-fvm9g\") pod \"openstack-operator-index-t8gtv\" (UID: \"8fd52a2a-b1af-4528-a6ba-2e369fed8b99\") " pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:10 crc kubenswrapper[4824]: I0122 10:00:10.011480    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t8gtv"]
Jan 22 10:00:10 crc kubenswrapper[4824]: I0122 10:00:10.086855    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvm9g\" (UniqueName: \"kubernetes.io/projected/8fd52a2a-b1af-4528-a6ba-2e369fed8b99-kube-api-access-fvm9g\") pod \"openstack-operator-index-t8gtv\" (UID: \"8fd52a2a-b1af-4528-a6ba-2e369fed8b99\") " pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:10 crc kubenswrapper[4824]: I0122 10:00:10.105325    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvm9g\" (UniqueName: \"kubernetes.io/projected/8fd52a2a-b1af-4528-a6ba-2e369fed8b99-kube-api-access-fvm9g\") pod \"openstack-operator-index-t8gtv\" (UID: \"8fd52a2a-b1af-4528-a6ba-2e369fed8b99\") " pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:10 crc kubenswrapper[4824]: I0122 10:00:10.288578    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:10 crc kubenswrapper[4824]: I0122 10:00:10.683008    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t8gtv"]
Jan 22 10:00:10 crc kubenswrapper[4824]: W0122 10:00:10.690114    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fd52a2a_b1af_4528_a6ba_2e369fed8b99.slice/crio-60255067caccfe947a44837458342d200fae3d1e10219faeea3f0ae9f6664585 WatchSource:0}: Error finding container 60255067caccfe947a44837458342d200fae3d1e10219faeea3f0ae9f6664585: Status 404 returned error can't find the container with id 60255067caccfe947a44837458342d200fae3d1e10219faeea3f0ae9f6664585
Jan 22 10:00:11 crc kubenswrapper[4824]: I0122 10:00:11.049826    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t8gtv" event={"ID":"8fd52a2a-b1af-4528-a6ba-2e369fed8b99","Type":"ContainerStarted","Data":"60255067caccfe947a44837458342d200fae3d1e10219faeea3f0ae9f6664585"}
Jan 22 10:00:13 crc kubenswrapper[4824]: I0122 10:00:13.300984    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-t8gtv"]
Jan 22 10:00:13 crc kubenswrapper[4824]: I0122 10:00:13.912945    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-dlbcw"]
Jan 22 10:00:13 crc kubenswrapper[4824]: I0122 10:00:13.914501    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:13 crc kubenswrapper[4824]: I0122 10:00:13.924751    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dlbcw"]
Jan 22 10:00:13 crc kubenswrapper[4824]: I0122 10:00:13.935043    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdgsf\" (UniqueName: \"kubernetes.io/projected/b85b4390-3574-4e36-bc3a-c0b739646d91-kube-api-access-vdgsf\") pod \"openstack-operator-index-dlbcw\" (UID: \"b85b4390-3574-4e36-bc3a-c0b739646d91\") " pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:14 crc kubenswrapper[4824]: I0122 10:00:14.035774    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdgsf\" (UniqueName: \"kubernetes.io/projected/b85b4390-3574-4e36-bc3a-c0b739646d91-kube-api-access-vdgsf\") pod \"openstack-operator-index-dlbcw\" (UID: \"b85b4390-3574-4e36-bc3a-c0b739646d91\") " pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:14 crc kubenswrapper[4824]: I0122 10:00:14.064839    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdgsf\" (UniqueName: \"kubernetes.io/projected/b85b4390-3574-4e36-bc3a-c0b739646d91-kube-api-access-vdgsf\") pod \"openstack-operator-index-dlbcw\" (UID: \"b85b4390-3574-4e36-bc3a-c0b739646d91\") " pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:14 crc kubenswrapper[4824]: I0122 10:00:14.078101    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t8gtv" event={"ID":"8fd52a2a-b1af-4528-a6ba-2e369fed8b99","Type":"ContainerStarted","Data":"f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9"}
Jan 22 10:00:14 crc kubenswrapper[4824]: I0122 10:00:14.098751    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t8gtv" podStartSLOduration=2.796626212 podStartE2EDuration="5.098722258s" podCreationTimestamp="2026-01-22 10:00:09 +0000 UTC" firstStartedPulling="2026-01-22 10:00:10.692547597 +0000 UTC m=+1043.596024580" lastFinishedPulling="2026-01-22 10:00:12.994643633 +0000 UTC m=+1045.898120626" observedRunningTime="2026-01-22 10:00:14.095661052 +0000 UTC m=+1046.999138115" watchObservedRunningTime="2026-01-22 10:00:14.098722258 +0000 UTC m=+1047.002199291"
Jan 22 10:00:14 crc kubenswrapper[4824]: I0122 10:00:14.235491    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:14 crc kubenswrapper[4824]: I0122 10:00:14.657978    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-dlbcw"]
Jan 22 10:00:15 crc kubenswrapper[4824]: I0122 10:00:15.095676    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dlbcw" event={"ID":"b85b4390-3574-4e36-bc3a-c0b739646d91","Type":"ContainerStarted","Data":"7a2768f0cb93b43845cc2d8a79c2bdfa37492301972255cea887f77ba7d243e1"}
Jan 22 10:00:15 crc kubenswrapper[4824]: I0122 10:00:15.096070    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-t8gtv" podUID="8fd52a2a-b1af-4528-a6ba-2e369fed8b99" containerName="registry-server" containerID="cri-o://f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9" gracePeriod=2
Jan 22 10:00:15 crc kubenswrapper[4824]: I0122 10:00:15.521645    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:15 crc kubenswrapper[4824]: I0122 10:00:15.555205    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvm9g\" (UniqueName: \"kubernetes.io/projected/8fd52a2a-b1af-4528-a6ba-2e369fed8b99-kube-api-access-fvm9g\") pod \"8fd52a2a-b1af-4528-a6ba-2e369fed8b99\" (UID: \"8fd52a2a-b1af-4528-a6ba-2e369fed8b99\") "
Jan 22 10:00:15 crc kubenswrapper[4824]: I0122 10:00:15.560108    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd52a2a-b1af-4528-a6ba-2e369fed8b99-kube-api-access-fvm9g" (OuterVolumeSpecName: "kube-api-access-fvm9g") pod "8fd52a2a-b1af-4528-a6ba-2e369fed8b99" (UID: "8fd52a2a-b1af-4528-a6ba-2e369fed8b99"). InnerVolumeSpecName "kube-api-access-fvm9g". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:00:15 crc kubenswrapper[4824]: I0122 10:00:15.655818    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvm9g\" (UniqueName: \"kubernetes.io/projected/8fd52a2a-b1af-4528-a6ba-2e369fed8b99-kube-api-access-fvm9g\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.105107    4824 generic.go:334] "Generic (PLEG): container finished" podID="8fd52a2a-b1af-4528-a6ba-2e369fed8b99" containerID="f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9" exitCode=0
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.105166    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t8gtv"
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.105168    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t8gtv" event={"ID":"8fd52a2a-b1af-4528-a6ba-2e369fed8b99","Type":"ContainerDied","Data":"f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9"}
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.105237    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t8gtv" event={"ID":"8fd52a2a-b1af-4528-a6ba-2e369fed8b99","Type":"ContainerDied","Data":"60255067caccfe947a44837458342d200fae3d1e10219faeea3f0ae9f6664585"}
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.105290    4824 scope.go:117] "RemoveContainer" containerID="f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9"
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.138665    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-t8gtv"]
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.139233    4824 scope.go:117] "RemoveContainer" containerID="f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9"
Jan 22 10:00:16 crc kubenswrapper[4824]: E0122 10:00:16.139964    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9\": container with ID starting with f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9 not found: ID does not exist" containerID="f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9"
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.140029    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9"} err="failed to get container status \"f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9\": rpc error: code = NotFound desc = could not find container \"f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9\": container with ID starting with f01552d63b2220b70169c8196950c42bc265558db9afddb350082502c2d2a8b9 not found: ID does not exist"
Jan 22 10:00:16 crc kubenswrapper[4824]: I0122 10:00:16.143107    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-t8gtv"]
Jan 22 10:00:17 crc kubenswrapper[4824]: I0122 10:00:17.417937    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd52a2a-b1af-4528-a6ba-2e369fed8b99" path="/var/lib/kubelet/pods/8fd52a2a-b1af-4528-a6ba-2e369fed8b99/volumes"
Jan 22 10:00:18 crc kubenswrapper[4824]: I0122 10:00:18.124964    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-dlbcw" event={"ID":"b85b4390-3574-4e36-bc3a-c0b739646d91","Type":"ContainerStarted","Data":"0b71cf25dcc796c0ef84dc2bde6e62d708ce40c7829c283e8c4fd074a49baf96"}
Jan 22 10:00:18 crc kubenswrapper[4824]: I0122 10:00:18.148083    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-dlbcw" podStartSLOduration=2.108282617 podStartE2EDuration="5.148055595s" podCreationTimestamp="2026-01-22 10:00:13 +0000 UTC" firstStartedPulling="2026-01-22 10:00:14.668613184 +0000 UTC m=+1047.572090177" lastFinishedPulling="2026-01-22 10:00:17.708386162 +0000 UTC m=+1050.611863155" observedRunningTime="2026-01-22 10:00:18.142060017 +0000 UTC m=+1051.045537050" watchObservedRunningTime="2026-01-22 10:00:18.148055595 +0000 UTC m=+1051.051532608"
Jan 22 10:00:24 crc kubenswrapper[4824]: I0122 10:00:24.236685    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:24 crc kubenswrapper[4824]: I0122 10:00:24.237737    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:24 crc kubenswrapper[4824]: I0122 10:00:24.283951    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:25 crc kubenswrapper[4824]: I0122 10:00:25.204130    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-dlbcw"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.348211    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"]
Jan 22 10:00:32 crc kubenswrapper[4824]: E0122 10:00:32.348854    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd52a2a-b1af-4528-a6ba-2e369fed8b99" containerName="registry-server"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.348875    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd52a2a-b1af-4528-a6ba-2e369fed8b99" containerName="registry-server"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.349030    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd52a2a-b1af-4528-a6ba-2e369fed8b99" containerName="registry-server"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.350216    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.353790    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-n8dhw"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.362202    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"]
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.394121    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb5mt\" (UniqueName: \"kubernetes.io/projected/176eeb37-df5a-41b4-90b6-16b04c0c576c-kube-api-access-xb5mt\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.394247    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-bundle\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.394294    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-util\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.495716    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-bundle\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.495805    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-util\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.496222    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb5mt\" (UniqueName: \"kubernetes.io/projected/176eeb37-df5a-41b4-90b6-16b04c0c576c-kube-api-access-xb5mt\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.497401    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-util\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.497564    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-bundle\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.532491    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb5mt\" (UniqueName: \"kubernetes.io/projected/176eeb37-df5a-41b4-90b6-16b04c0c576c-kube-api-access-xb5mt\") pod \"fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") " pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:32 crc kubenswrapper[4824]: I0122 10:00:32.693114    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:33 crc kubenswrapper[4824]: I0122 10:00:33.137472    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"]
Jan 22 10:00:33 crc kubenswrapper[4824]: I0122 10:00:33.228570    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt" event={"ID":"176eeb37-df5a-41b4-90b6-16b04c0c576c","Type":"ContainerStarted","Data":"58b935e1f89f33eab77f786912aae1a15ccc97fcdc518a19a61af1ac1feaebee"}
Jan 22 10:00:35 crc kubenswrapper[4824]: I0122 10:00:35.244744    4824 generic.go:334] "Generic (PLEG): container finished" podID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerID="345a1eb9c91279a0af04e63de718f63c9e3b131e30abb2168b7c0accbbedf945" exitCode=0
Jan 22 10:00:35 crc kubenswrapper[4824]: I0122 10:00:35.245040    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt" event={"ID":"176eeb37-df5a-41b4-90b6-16b04c0c576c","Type":"ContainerDied","Data":"345a1eb9c91279a0af04e63de718f63c9e3b131e30abb2168b7c0accbbedf945"}
Jan 22 10:00:37 crc kubenswrapper[4824]: I0122 10:00:37.260803    4824 generic.go:334] "Generic (PLEG): container finished" podID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerID="63170b9e827f539deb3bdf35784438f3d5f392655911befd4098d8d8c093be2f" exitCode=0
Jan 22 10:00:37 crc kubenswrapper[4824]: I0122 10:00:37.260892    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt" event={"ID":"176eeb37-df5a-41b4-90b6-16b04c0c576c","Type":"ContainerDied","Data":"63170b9e827f539deb3bdf35784438f3d5f392655911befd4098d8d8c093be2f"}
Jan 22 10:00:38 crc kubenswrapper[4824]: I0122 10:00:38.272074    4824 generic.go:334] "Generic (PLEG): container finished" podID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerID="66075ccd3574e6a92057a665ade8eb08d322b8b215838fafdd2a0fd9bc90ef5a" exitCode=0
Jan 22 10:00:38 crc kubenswrapper[4824]: I0122 10:00:38.272147    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt" event={"ID":"176eeb37-df5a-41b4-90b6-16b04c0c576c","Type":"ContainerDied","Data":"66075ccd3574e6a92057a665ade8eb08d322b8b215838fafdd2a0fd9bc90ef5a"}
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.532952    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.687119    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb5mt\" (UniqueName: \"kubernetes.io/projected/176eeb37-df5a-41b4-90b6-16b04c0c576c-kube-api-access-xb5mt\") pod \"176eeb37-df5a-41b4-90b6-16b04c0c576c\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") "
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.687278    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-bundle\") pod \"176eeb37-df5a-41b4-90b6-16b04c0c576c\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") "
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.687313    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-util\") pod \"176eeb37-df5a-41b4-90b6-16b04c0c576c\" (UID: \"176eeb37-df5a-41b4-90b6-16b04c0c576c\") "
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.687965    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-bundle" (OuterVolumeSpecName: "bundle") pod "176eeb37-df5a-41b4-90b6-16b04c0c576c" (UID: "176eeb37-df5a-41b4-90b6-16b04c0c576c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.699611    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/176eeb37-df5a-41b4-90b6-16b04c0c576c-kube-api-access-xb5mt" (OuterVolumeSpecName: "kube-api-access-xb5mt") pod "176eeb37-df5a-41b4-90b6-16b04c0c576c" (UID: "176eeb37-df5a-41b4-90b6-16b04c0c576c"). InnerVolumeSpecName "kube-api-access-xb5mt". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.700405    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-util" (OuterVolumeSpecName: "util") pod "176eeb37-df5a-41b4-90b6-16b04c0c576c" (UID: "176eeb37-df5a-41b4-90b6-16b04c0c576c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.789279    4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-bundle\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.789332    4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/176eeb37-df5a-41b4-90b6-16b04c0c576c-util\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:39 crc kubenswrapper[4824]: I0122 10:00:39.789352    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb5mt\" (UniqueName: \"kubernetes.io/projected/176eeb37-df5a-41b4-90b6-16b04c0c576c-kube-api-access-xb5mt\") on node \"crc\" DevicePath \"\""
Jan 22 10:00:40 crc kubenswrapper[4824]: I0122 10:00:40.289034    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt" event={"ID":"176eeb37-df5a-41b4-90b6-16b04c0c576c","Type":"ContainerDied","Data":"58b935e1f89f33eab77f786912aae1a15ccc97fcdc518a19a61af1ac1feaebee"}
Jan 22 10:00:40 crc kubenswrapper[4824]: I0122 10:00:40.289119    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b935e1f89f33eab77f786912aae1a15ccc97fcdc518a19a61af1ac1feaebee"
Jan 22 10:00:40 crc kubenswrapper[4824]: I0122 10:00:40.289075    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.547743    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"]
Jan 22 10:00:44 crc kubenswrapper[4824]: E0122 10:00:44.548492    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="util"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.548507    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="util"
Jan 22 10:00:44 crc kubenswrapper[4824]: E0122 10:00:44.548525    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="extract"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.548534    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="extract"
Jan 22 10:00:44 crc kubenswrapper[4824]: E0122 10:00:44.548549    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="pull"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.548558    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="pull"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.548690    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="176eeb37-df5a-41b4-90b6-16b04c0c576c" containerName="extract"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.549190    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.552930    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-8zs6c"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.639564    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"]
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.648692    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz56f\" (UniqueName: \"kubernetes.io/projected/ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4-kube-api-access-hz56f\") pod \"openstack-operator-controller-init-698d6bb84b-mv4kv\" (UID: \"ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4\") " pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.750031    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz56f\" (UniqueName: \"kubernetes.io/projected/ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4-kube-api-access-hz56f\") pod \"openstack-operator-controller-init-698d6bb84b-mv4kv\" (UID: \"ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4\") " pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.769094    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz56f\" (UniqueName: \"kubernetes.io/projected/ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4-kube-api-access-hz56f\") pod \"openstack-operator-controller-init-698d6bb84b-mv4kv\" (UID: \"ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4\") " pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:00:44 crc kubenswrapper[4824]: I0122 10:00:44.875059    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:00:45 crc kubenswrapper[4824]: I0122 10:00:45.304459    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"]
Jan 22 10:00:45 crc kubenswrapper[4824]: I0122 10:00:45.322489    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv" event={"ID":"ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4","Type":"ContainerStarted","Data":"3be31576ee29c921bc1010d2df3bf92d856e88222df5e41d836ba2e1b8247593"}
Jan 22 10:00:50 crc kubenswrapper[4824]: I0122 10:00:50.356243    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv" event={"ID":"ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4","Type":"ContainerStarted","Data":"f154f13d02d7e599f8e005c4fc1707d32cba93243f1f4c2a32bfefb619605d0a"}
Jan 22 10:00:50 crc kubenswrapper[4824]: I0122 10:00:50.357489    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:00:50 crc kubenswrapper[4824]: I0122 10:00:50.402077    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv" podStartSLOduration=1.963536199 podStartE2EDuration="6.402050436s" podCreationTimestamp="2026-01-22 10:00:44 +0000 UTC" firstStartedPulling="2026-01-22 10:00:45.318692654 +0000 UTC m=+1078.222169647" lastFinishedPulling="2026-01-22 10:00:49.757206891 +0000 UTC m=+1082.660683884" observedRunningTime="2026-01-22 10:00:50.39504401 +0000 UTC m=+1083.298521013" watchObservedRunningTime="2026-01-22 10:00:50.402050436 +0000 UTC m=+1083.305527469"
Jan 22 10:00:54 crc kubenswrapper[4824]: I0122 10:00:54.878242    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-698d6bb84b-mv4kv"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.643154    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.645446    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.647481    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-bvgww"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.649126    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.649872    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.652776    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qb88w"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.663215    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.670244    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.671162    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.674770    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2sj7j"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.689420    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.705899    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.716419    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.717397    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.727219    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-98dtx"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.739541    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.760187    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.760948    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.764884    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ccnxw"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.800706    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5d8k\" (UniqueName: \"kubernetes.io/projected/905ddc8e-1123-4654-92a3-a50493b92a78-kube-api-access-t5d8k\") pod \"designate-operator-controller-manager-b45d7bf98-8prhs\" (UID: \"905ddc8e-1123-4654-92a3-a50493b92a78\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.800818    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkksm\" (UniqueName: \"kubernetes.io/projected/b53e7faf-8d7c-4f40-82c7-e6aa64508488-kube-api-access-hkksm\") pod \"glance-operator-controller-manager-78fdd796fd-rt6zn\" (UID: \"b53e7faf-8d7c-4f40-82c7-e6aa64508488\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.800883    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpt6d\" (UniqueName: \"kubernetes.io/projected/88e0e1fa-e0ae-43e1-b653-158ea72244b1-kube-api-access-mpt6d\") pod \"barbican-operator-controller-manager-59dd8b7cbf-nbrrg\" (UID: \"88e0e1fa-e0ae-43e1-b653-158ea72244b1\") " pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.800933    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gccb4\" (UniqueName: \"kubernetes.io/projected/d29a7a0c-344a-449a-bfa0-457e860f9030-kube-api-access-gccb4\") pod \"cinder-operator-controller-manager-69cf5d4557-vshwt\" (UID: \"d29a7a0c-344a-449a-bfa0-457e860f9030\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.844042    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.871698    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.872921    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.875963    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-bbh97"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.881187    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.882188    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.889127    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jhzqr"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.889252    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.892417    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.904771    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5d8k\" (UniqueName: \"kubernetes.io/projected/905ddc8e-1123-4654-92a3-a50493b92a78-kube-api-access-t5d8k\") pod \"designate-operator-controller-manager-b45d7bf98-8prhs\" (UID: \"905ddc8e-1123-4654-92a3-a50493b92a78\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.904819    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkksm\" (UniqueName: \"kubernetes.io/projected/b53e7faf-8d7c-4f40-82c7-e6aa64508488-kube-api-access-hkksm\") pod \"glance-operator-controller-manager-78fdd796fd-rt6zn\" (UID: \"b53e7faf-8d7c-4f40-82c7-e6aa64508488\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.904851    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpt6d\" (UniqueName: \"kubernetes.io/projected/88e0e1fa-e0ae-43e1-b653-158ea72244b1-kube-api-access-mpt6d\") pod \"barbican-operator-controller-manager-59dd8b7cbf-nbrrg\" (UID: \"88e0e1fa-e0ae-43e1-b653-158ea72244b1\") " pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.904880    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp6sj\" (UniqueName: \"kubernetes.io/projected/792e3f5e-1236-4fe0-94b5-20e322aae84c-kube-api-access-wp6sj\") pod \"heat-operator-controller-manager-594c8c9d5d-f5qrl\" (UID: \"792e3f5e-1236-4fe0-94b5-20e322aae84c\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.904901    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gccb4\" (UniqueName: \"kubernetes.io/projected/d29a7a0c-344a-449a-bfa0-457e860f9030-kube-api-access-gccb4\") pod \"cinder-operator-controller-manager-69cf5d4557-vshwt\" (UID: \"d29a7a0c-344a-449a-bfa0-457e860f9030\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.932139    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.933420    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.940772    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.942777    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-pz2qs"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.947386    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpt6d\" (UniqueName: \"kubernetes.io/projected/88e0e1fa-e0ae-43e1-b653-158ea72244b1-kube-api-access-mpt6d\") pod \"barbican-operator-controller-manager-59dd8b7cbf-nbrrg\" (UID: \"88e0e1fa-e0ae-43e1-b653-158ea72244b1\") " pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.949125    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5d8k\" (UniqueName: \"kubernetes.io/projected/905ddc8e-1123-4654-92a3-a50493b92a78-kube-api-access-t5d8k\") pod \"designate-operator-controller-manager-b45d7bf98-8prhs\" (UID: \"905ddc8e-1123-4654-92a3-a50493b92a78\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.950194    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gccb4\" (UniqueName: \"kubernetes.io/projected/d29a7a0c-344a-449a-bfa0-457e860f9030-kube-api-access-gccb4\") pod \"cinder-operator-controller-manager-69cf5d4557-vshwt\" (UID: \"d29a7a0c-344a-449a-bfa0-457e860f9030\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.959091    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.960013    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.965200    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"]
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.965683    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-cdj6q"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.965879    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.976582    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.987585    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkksm\" (UniqueName: \"kubernetes.io/projected/b53e7faf-8d7c-4f40-82c7-e6aa64508488-kube-api-access-hkksm\") pod \"glance-operator-controller-manager-78fdd796fd-rt6zn\" (UID: \"b53e7faf-8d7c-4f40-82c7-e6aa64508488\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:01:33 crc kubenswrapper[4824]: I0122 10:01:33.993475    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.001330    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.006854    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg76n\" (UniqueName: \"kubernetes.io/projected/6507116d-ad87-4be1-8322-340170f575b3-kube-api-access-wg76n\") pod \"horizon-operator-controller-manager-77d5c5b54f-k2hwm\" (UID: \"6507116d-ad87-4be1-8322-340170f575b3\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.006900    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.006948    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp6sj\" (UniqueName: \"kubernetes.io/projected/792e3f5e-1236-4fe0-94b5-20e322aae84c-kube-api-access-wp6sj\") pod \"heat-operator-controller-manager-594c8c9d5d-f5qrl\" (UID: \"792e3f5e-1236-4fe0-94b5-20e322aae84c\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.007003    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwch5\" (UniqueName: \"kubernetes.io/projected/40210948-0e80-4f69-9d75-ab44c0d38285-kube-api-access-rwch5\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.007028    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td8pz\" (UniqueName: \"kubernetes.io/projected/2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9-kube-api-access-td8pz\") pod \"ironic-operator-controller-manager-69d6c9f5b8-npvrh\" (UID: \"2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9\") " pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.020481    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.021601    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.025175    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.027254    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-7pqzl"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.043432    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp6sj\" (UniqueName: \"kubernetes.io/projected/792e3f5e-1236-4fe0-94b5-20e322aae84c-kube-api-access-wp6sj\") pod \"heat-operator-controller-manager-594c8c9d5d-f5qrl\" (UID: \"792e3f5e-1236-4fe0-94b5-20e322aae84c\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.046145    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.049634    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.050578    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.052438    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mqzkx"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.057625    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.057754    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.063266    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-6vmzd"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.065150    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.065341    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.067407    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.073195    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.074877    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-8j2zn"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.085399    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.090676    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.092188    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.095665    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-bqgk4"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.102407    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.108479    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.108569    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rn6m\" (UniqueName: \"kubernetes.io/projected/f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93-kube-api-access-8rn6m\") pod \"manila-operator-controller-manager-78c6999f6f-zwfpg\" (UID: \"f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.108615    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z78c\" (UniqueName: \"kubernetes.io/projected/1b0c67df-ba69-46b3-ad14-e7287f34fd27-kube-api-access-9z78c\") pod \"keystone-operator-controller-manager-b8b6d4659-jkg9n\" (UID: \"1b0c67df-ba69-46b3-ad14-e7287f34fd27\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.108635    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwch5\" (UniqueName: \"kubernetes.io/projected/40210948-0e80-4f69-9d75-ab44c0d38285-kube-api-access-rwch5\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.108662    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td8pz\" (UniqueName: \"kubernetes.io/projected/2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9-kube-api-access-td8pz\") pod \"ironic-operator-controller-manager-69d6c9f5b8-npvrh\" (UID: \"2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9\") " pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.108698    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg76n\" (UniqueName: \"kubernetes.io/projected/6507116d-ad87-4be1-8322-340170f575b3-kube-api-access-wg76n\") pod \"horizon-operator-controller-manager-77d5c5b54f-k2hwm\" (UID: \"6507116d-ad87-4be1-8322-340170f575b3\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.109667    4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.109754    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert podName:40210948-0e80-4f69-9d75-ab44c0d38285 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:34.609721475 +0000 UTC m=+1127.513198468 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert") pod "infra-operator-controller-manager-54ccf4f85d-gbhz2" (UID: "40210948-0e80-4f69-9d75-ab44c0d38285") : secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.111050    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.141622    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.142578    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.155453    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.155729    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mbmxr"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.158243    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg76n\" (UniqueName: \"kubernetes.io/projected/6507116d-ad87-4be1-8322-340170f575b3-kube-api-access-wg76n\") pod \"horizon-operator-controller-manager-77d5c5b54f-k2hwm\" (UID: \"6507116d-ad87-4be1-8322-340170f575b3\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.163814    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwch5\" (UniqueName: \"kubernetes.io/projected/40210948-0e80-4f69-9d75-ab44c0d38285-kube-api-access-rwch5\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.168472    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td8pz\" (UniqueName: \"kubernetes.io/projected/2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9-kube-api-access-td8pz\") pod \"ironic-operator-controller-manager-69d6c9f5b8-npvrh\" (UID: \"2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9\") " pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.174649    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213347    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pv2z\" (UniqueName: \"kubernetes.io/projected/59abaeff-0c1b-4f9d-bd55-448a1295f603-kube-api-access-4pv2z\") pod \"mariadb-operator-controller-manager-c87fff755-krm9d\" (UID: \"59abaeff-0c1b-4f9d-bd55-448a1295f603\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213420    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd5w7\" (UniqueName: \"kubernetes.io/projected/aa122c78-b0a4-4bd5-9844-7a44b150dac4-kube-api-access-fd5w7\") pod \"nova-operator-controller-manager-6b8bc8d87d-hrzwc\" (UID: \"aa122c78-b0a4-4bd5-9844-7a44b150dac4\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213474    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rn6m\" (UniqueName: \"kubernetes.io/projected/f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93-kube-api-access-8rn6m\") pod \"manila-operator-controller-manager-78c6999f6f-zwfpg\" (UID: \"f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213504    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz94p\" (UniqueName: \"kubernetes.io/projected/d5b9d0a8-b7a8-4138-a306-4599159f33e1-kube-api-access-sz94p\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213543    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213575    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z78c\" (UniqueName: \"kubernetes.io/projected/1b0c67df-ba69-46b3-ad14-e7287f34fd27-kube-api-access-9z78c\") pod \"keystone-operator-controller-manager-b8b6d4659-jkg9n\" (UID: \"1b0c67df-ba69-46b3-ad14-e7287f34fd27\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213652    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8p7h\" (UniqueName: \"kubernetes.io/projected/651ec3cd-3884-41c6-ae9d-1fb08b764a32-kube-api-access-s8p7h\") pod \"neutron-operator-controller-manager-5d8f59fb49-q25f6\" (UID: \"651ec3cd-3884-41c6-ae9d-1fb08b764a32\") " pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.213678    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbgn4\" (UniqueName: \"kubernetes.io/projected/cb110aec-f41e-4bae-b0de-9c076a41a603-kube-api-access-dbgn4\") pod \"octavia-operator-controller-manager-7bd9774b6-wm2vt\" (UID: \"cb110aec-f41e-4bae-b0de-9c076a41a603\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.215053    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.234981    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.236304    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.241682    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-9b6w4"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.257163    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rn6m\" (UniqueName: \"kubernetes.io/projected/f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93-kube-api-access-8rn6m\") pod \"manila-operator-controller-manager-78c6999f6f-zwfpg\" (UID: \"f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.260435    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.261761    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.273810    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2vk87"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.283952    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z78c\" (UniqueName: \"kubernetes.io/projected/1b0c67df-ba69-46b3-ad14-e7287f34fd27-kube-api-access-9z78c\") pod \"keystone-operator-controller-manager-b8b6d4659-jkg9n\" (UID: \"1b0c67df-ba69-46b3-ad14-e7287f34fd27\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.293428    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.294415    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.297930    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.306971    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-h89vj"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.307184    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.315241    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.318186    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxnjd\" (UniqueName: \"kubernetes.io/projected/2a362e5c-b09f-4c6c-8639-6c7107e049b0-kube-api-access-gxnjd\") pod \"placement-operator-controller-manager-5d646b7d76-vgkz5\" (UID: \"2a362e5c-b09f-4c6c-8639-6c7107e049b0\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.318260    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz94p\" (UniqueName: \"kubernetes.io/projected/d5b9d0a8-b7a8-4138-a306-4599159f33e1-kube-api-access-sz94p\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.318314    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.318446    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8p7h\" (UniqueName: \"kubernetes.io/projected/651ec3cd-3884-41c6-ae9d-1fb08b764a32-kube-api-access-s8p7h\") pod \"neutron-operator-controller-manager-5d8f59fb49-q25f6\" (UID: \"651ec3cd-3884-41c6-ae9d-1fb08b764a32\") " pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.318483    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbgn4\" (UniqueName: \"kubernetes.io/projected/cb110aec-f41e-4bae-b0de-9c076a41a603-kube-api-access-dbgn4\") pod \"octavia-operator-controller-manager-7bd9774b6-wm2vt\" (UID: \"cb110aec-f41e-4bae-b0de-9c076a41a603\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.319944    4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.343258    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert podName:d5b9d0a8-b7a8-4138-a306-4599159f33e1 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:34.843203661 +0000 UTC m=+1127.746680654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" (UID: "d5b9d0a8-b7a8-4138-a306-4599159f33e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.320022    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pv2z\" (UniqueName: \"kubernetes.io/projected/59abaeff-0c1b-4f9d-bd55-448a1295f603-kube-api-access-4pv2z\") pod \"mariadb-operator-controller-manager-c87fff755-krm9d\" (UID: \"59abaeff-0c1b-4f9d-bd55-448a1295f603\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.343710    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd5w7\" (UniqueName: \"kubernetes.io/projected/aa122c78-b0a4-4bd5-9844-7a44b150dac4-kube-api-access-fd5w7\") pod \"nova-operator-controller-manager-6b8bc8d87d-hrzwc\" (UID: \"aa122c78-b0a4-4bd5-9844-7a44b150dac4\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.364347    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.375089    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz94p\" (UniqueName: \"kubernetes.io/projected/d5b9d0a8-b7a8-4138-a306-4599159f33e1-kube-api-access-sz94p\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.381487    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pv2z\" (UniqueName: \"kubernetes.io/projected/59abaeff-0c1b-4f9d-bd55-448a1295f603-kube-api-access-4pv2z\") pod \"mariadb-operator-controller-manager-c87fff755-krm9d\" (UID: \"59abaeff-0c1b-4f9d-bd55-448a1295f603\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.382373    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbgn4\" (UniqueName: \"kubernetes.io/projected/cb110aec-f41e-4bae-b0de-9c076a41a603-kube-api-access-dbgn4\") pod \"octavia-operator-controller-manager-7bd9774b6-wm2vt\" (UID: \"cb110aec-f41e-4bae-b0de-9c076a41a603\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.384875    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd5w7\" (UniqueName: \"kubernetes.io/projected/aa122c78-b0a4-4bd5-9844-7a44b150dac4-kube-api-access-fd5w7\") pod \"nova-operator-controller-manager-6b8bc8d87d-hrzwc\" (UID: \"aa122c78-b0a4-4bd5-9844-7a44b150dac4\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.386984    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.412879    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.439514    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.448110    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxnjd\" (UniqueName: \"kubernetes.io/projected/2a362e5c-b09f-4c6c-8639-6c7107e049b0-kube-api-access-gxnjd\") pod \"placement-operator-controller-manager-5d646b7d76-vgkz5\" (UID: \"2a362e5c-b09f-4c6c-8639-6c7107e049b0\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.448338    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxk9m\" (UniqueName: \"kubernetes.io/projected/3370a46a-f9bf-4ce2-977b-6e910b3a4448-kube-api-access-fxk9m\") pod \"swift-operator-controller-manager-547cbdb99f-ch89p\" (UID: \"3370a46a-f9bf-4ce2-977b-6e910b3a4448\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.448439    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l84rg\" (UniqueName: \"kubernetes.io/projected/0cf0de8e-51d5-4f14-a63b-08acab6d68a7-kube-api-access-l84rg\") pod \"ovn-operator-controller-manager-55db956ddc-l8kpr\" (UID: \"0cf0de8e-51d5-4f14-a63b-08acab6d68a7\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.448675    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.448718    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.472233    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-m6j66"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.485790    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8p7h\" (UniqueName: \"kubernetes.io/projected/651ec3cd-3884-41c6-ae9d-1fb08b764a32-kube-api-access-s8p7h\") pod \"neutron-operator-controller-manager-5d8f59fb49-q25f6\" (UID: \"651ec3cd-3884-41c6-ae9d-1fb08b764a32\") " pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.487641    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.490212    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.504438    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxnjd\" (UniqueName: \"kubernetes.io/projected/2a362e5c-b09f-4c6c-8639-6c7107e049b0-kube-api-access-gxnjd\") pod \"placement-operator-controller-manager-5d646b7d76-vgkz5\" (UID: \"2a362e5c-b09f-4c6c-8639-6c7107e049b0\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.516312    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.546027    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.552148    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxk9m\" (UniqueName: \"kubernetes.io/projected/3370a46a-f9bf-4ce2-977b-6e910b3a4448-kube-api-access-fxk9m\") pod \"swift-operator-controller-manager-547cbdb99f-ch89p\" (UID: \"3370a46a-f9bf-4ce2-977b-6e910b3a4448\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.552251    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7ckl\" (UniqueName: \"kubernetes.io/projected/2b9855d9-90b7-4904-9a65-a06a4674d86e-kube-api-access-q7ckl\") pod \"telemetry-operator-controller-manager-85cd9769bb-rcxtb\" (UID: \"2b9855d9-90b7-4904-9a65-a06a4674d86e\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.552288    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l84rg\" (UniqueName: \"kubernetes.io/projected/0cf0de8e-51d5-4f14-a63b-08acab6d68a7-kube-api-access-l84rg\") pod \"ovn-operator-controller-manager-55db956ddc-l8kpr\" (UID: \"0cf0de8e-51d5-4f14-a63b-08acab6d68a7\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.571948    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.572776    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.572860    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.588193    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-l52wh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.591067    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxk9m\" (UniqueName: \"kubernetes.io/projected/3370a46a-f9bf-4ce2-977b-6e910b3a4448-kube-api-access-fxk9m\") pod \"swift-operator-controller-manager-547cbdb99f-ch89p\" (UID: \"3370a46a-f9bf-4ce2-977b-6e910b3a4448\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.597960    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.598930    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.601456    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-464vf"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.607116    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l84rg\" (UniqueName: \"kubernetes.io/projected/0cf0de8e-51d5-4f14-a63b-08acab6d68a7-kube-api-access-l84rg\") pod \"ovn-operator-controller-manager-55db956ddc-l8kpr\" (UID: \"0cf0de8e-51d5-4f14-a63b-08acab6d68a7\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.621795    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.652973    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.654672    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.656530    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbczv\" (UniqueName: \"kubernetes.io/projected/14b5c974-8804-44d3-9b48-f28f0acab0b6-kube-api-access-mbczv\") pod \"test-operator-controller-manager-69797bbcbd-7wxgh\" (UID: \"14b5c974-8804-44d3-9b48-f28f0acab0b6\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.656676    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7ckl\" (UniqueName: \"kubernetes.io/projected/2b9855d9-90b7-4904-9a65-a06a4674d86e-kube-api-access-q7ckl\") pod \"telemetry-operator-controller-manager-85cd9769bb-rcxtb\" (UID: \"2b9855d9-90b7-4904-9a65-a06a4674d86e\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.657101    4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.657163    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert podName:40210948-0e80-4f69-9d75-ab44c0d38285 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:35.657144593 +0000 UTC m=+1128.560621586 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert") pod "infra-operator-controller-manager-54ccf4f85d-gbhz2" (UID: "40210948-0e80-4f69-9d75-ab44c0d38285") : secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.657377    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.659664    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.660071    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.660083    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sphf7"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.665487    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.672650    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.673768    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.676051    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-l8xlv"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.679547    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.680312    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7ckl\" (UniqueName: \"kubernetes.io/projected/2b9855d9-90b7-4904-9a65-a06a4674d86e-kube-api-access-q7ckl\") pod \"telemetry-operator-controller-manager-85cd9769bb-rcxtb\" (UID: \"2b9855d9-90b7-4904-9a65-a06a4674d86e\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.683171    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.736751    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.757210    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.758339    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.758430    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbczv\" (UniqueName: \"kubernetes.io/projected/14b5c974-8804-44d3-9b48-f28f0acab0b6-kube-api-access-mbczv\") pod \"test-operator-controller-manager-69797bbcbd-7wxgh\" (UID: \"14b5c974-8804-44d3-9b48-f28f0acab0b6\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.758465    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.758496    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5zjl\" (UniqueName: \"kubernetes.io/projected/34941af2-d22e-4aaf-b1b6-a620cfc975be-kube-api-access-x5zjl\") pod \"watcher-operator-controller-manager-5ffb9c6597-84f2h\" (UID: \"34941af2-d22e-4aaf-b1b6-a620cfc975be\") " pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.758537    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtmmf\" (UniqueName: \"kubernetes.io/projected/18d7f488-bb97-4a72-8fbc-eb5db501db16-kube-api-access-wtmmf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mhlsp\" (UID: \"18d7f488-bb97-4a72-8fbc-eb5db501db16\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.758567    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hkkr\" (UniqueName: \"kubernetes.io/projected/12e59f22-9370-4b23-8d8a-5c24ba925541-kube-api-access-4hkkr\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.783761    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbczv\" (UniqueName: \"kubernetes.io/projected/14b5c974-8804-44d3-9b48-f28f0acab0b6-kube-api-access-mbczv\") pod \"test-operator-controller-manager-69797bbcbd-7wxgh\" (UID: \"14b5c974-8804-44d3-9b48-f28f0acab0b6\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.838382    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.862646    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.862708    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.862738    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.863175    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5zjl\" (UniqueName: \"kubernetes.io/projected/34941af2-d22e-4aaf-b1b6-a620cfc975be-kube-api-access-x5zjl\") pod \"watcher-operator-controller-manager-5ffb9c6597-84f2h\" (UID: \"34941af2-d22e-4aaf-b1b6-a620cfc975be\") " pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.863264    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtmmf\" (UniqueName: \"kubernetes.io/projected/18d7f488-bb97-4a72-8fbc-eb5db501db16-kube-api-access-wtmmf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mhlsp\" (UID: \"18d7f488-bb97-4a72-8fbc-eb5db501db16\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.863312    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hkkr\" (UniqueName: \"kubernetes.io/projected/12e59f22-9370-4b23-8d8a-5c24ba925541-kube-api-access-4hkkr\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.863807    4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.863883    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:35.363862504 +0000 UTC m=+1128.267339497 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.863956    4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.864004    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert podName:d5b9d0a8-b7a8-4138-a306-4599159f33e1 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:35.863985368 +0000 UTC m=+1128.767462361 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" (UID: "d5b9d0a8-b7a8-4138-a306-4599159f33e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.864041    4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: E0122 10:01:34.864060    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:35.36405399 +0000 UTC m=+1128.267530983 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "metrics-server-cert" not found
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.881165    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5zjl\" (UniqueName: \"kubernetes.io/projected/34941af2-d22e-4aaf-b1b6-a620cfc975be-kube-api-access-x5zjl\") pod \"watcher-operator-controller-manager-5ffb9c6597-84f2h\" (UID: \"34941af2-d22e-4aaf-b1b6-a620cfc975be\") " pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.884801    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtmmf\" (UniqueName: \"kubernetes.io/projected/18d7f488-bb97-4a72-8fbc-eb5db501db16-kube-api-access-wtmmf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mhlsp\" (UID: \"18d7f488-bb97-4a72-8fbc-eb5db501db16\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.888782    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hkkr\" (UniqueName: \"kubernetes.io/projected/12e59f22-9370-4b23-8d8a-5c24ba925541-kube-api-access-4hkkr\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.909660    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"]
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.940236    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:01:34 crc kubenswrapper[4824]: I0122 10:01:34.950296    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.060446    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.064791    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.382240    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.382332    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.382505    4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.382564    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:36.382543283 +0000 UTC m=+1129.286020276 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "webhook-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.382925    4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.382961    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:36.382951054 +0000 UTC m=+1129.286428047 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "metrics-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.394627    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.417640    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.582936    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.602238    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.621223    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.680855    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn" event={"ID":"b53e7faf-8d7c-4f40-82c7-e6aa64508488","Type":"ContainerStarted","Data":"e47b314a5072151f948ea617488fcc2ebe545e8f8ad838a48de7ab768287a81b"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.682026    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n" event={"ID":"1b0c67df-ba69-46b3-ad14-e7287f34fd27","Type":"ContainerStarted","Data":"97445fd3a87a89400f4daa92d556a4c9201bbb0d2184aa5b8f2686933e7de475"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.682879    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm" event={"ID":"6507116d-ad87-4be1-8322-340170f575b3","Type":"ContainerStarted","Data":"bc2ce11f7dba80de6c23c4ac50acbbe917d2a3067e75914ed6906c49ce12245c"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.683682    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg" event={"ID":"88e0e1fa-e0ae-43e1-b653-158ea72244b1","Type":"ContainerStarted","Data":"1b024353235688f5a9e64ed748176e72e1b0c2e09080dd3cab4c42e66d7d513f"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.686088    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.686273    4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.686319    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert podName:40210948-0e80-4f69-9d75-ab44c0d38285 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:37.686302695 +0000 UTC m=+1130.589779688 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert") pod "infra-operator-controller-manager-54ccf4f85d-gbhz2" (UID: "40210948-0e80-4f69-9d75-ab44c0d38285") : secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.686697    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs" event={"ID":"905ddc8e-1123-4654-92a3-a50493b92a78","Type":"ContainerStarted","Data":"1128113c2d88f4a0b2c786ce2e4c733c8f449302f9ab8b1c2b6299de06de7ae4"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.687854    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl" event={"ID":"792e3f5e-1236-4fe0-94b5-20e322aae84c","Type":"ContainerStarted","Data":"babdabfd3da3ec729f4e55c7c838b6d520a4725c9642f9bf66d3610708b16863"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.688742    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt" event={"ID":"d29a7a0c-344a-449a-bfa0-457e860f9030","Type":"ContainerStarted","Data":"f5c38c0026e7da2960982e0e75aed84ed7fa447dcb1fa4a72433e45d2fe96434"}
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.805469    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.812566    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.828573    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"]
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.838601    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa122c78_b0a4_4bd5_9844_7a44b150dac4.slice/crio-f23e146657938b961c8e2212ec5852c22d771338e3995d6ebfc890f8c93f2efc WatchSource:0}: Error finding container f23e146657938b961c8e2212ec5852c22d771338e3995d6ebfc890f8c93f2efc: Status 404 returned error can't find the container with id f23e146657938b961c8e2212ec5852c22d771338e3995d6ebfc890f8c93f2efc
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.850296    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.860489    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.870798    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.885472    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.889595    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.889786    4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.889857    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert podName:d5b9d0a8-b7a8-4138-a306-4599159f33e1 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:37.889838996 +0000 UTC m=+1130.793315989 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" (UID: "d5b9d0a8-b7a8-4138-a306-4599159f33e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.892432    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod651ec3cd_3884_41c6_ae9d_1fb08b764a32.slice/crio-4c6532eac33f970c230cb31afe56a634506f20c58be3cbdfaac3d921197cede9 WatchSource:0}: Error finding container 4c6532eac33f970c230cb31afe56a634506f20c58be3cbdfaac3d921197cede9: Status 404 returned error can't find the container with id 4c6532eac33f970c230cb31afe56a634506f20c58be3cbdfaac3d921197cede9
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.899592    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.904463    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.911876    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.921259    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.927409    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp"]
Jan 22 10:01:35 crc kubenswrapper[4824]: I0122 10:01:35.932043    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"]
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.939492    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59abaeff_0c1b_4f9d_bd55_448a1295f603.slice/crio-2ddddd679203698cb9fae800c0cff0851a1f2a23eac31cdd3905d1dcf672684e WatchSource:0}: Error finding container 2ddddd679203698cb9fae800c0cff0851a1f2a23eac31cdd3905d1dcf672684e: Status 404 returned error can't find the container with id 2ddddd679203698cb9fae800c0cff0851a1f2a23eac31cdd3905d1dcf672684e
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.942390    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14b5c974_8804_44d3_9b48_f28f0acab0b6.slice/crio-edf8c4a65fe4df94481aec815cdc797b16e230e3148aae64672c016b84cce240 WatchSource:0}: Error finding container edf8c4a65fe4df94481aec815cdc797b16e230e3148aae64672c016b84cce240: Status 404 returned error can't find the container with id edf8c4a65fe4df94481aec815cdc797b16e230e3148aae64672c016b84cce240
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.943291    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b9855d9_90b7_4904_9a65_a06a4674d86e.slice/crio-4ef51980a141924c38c1e915e1b024c7ff98c5f20fdc2c44315f774a8af7f30a WatchSource:0}: Error finding container 4ef51980a141924c38c1e915e1b024c7ff98c5f20fdc2c44315f774a8af7f30a: Status 404 returned error can't find the container with id 4ef51980a141924c38c1e915e1b024c7ff98c5f20fdc2c44315f774a8af7f30a
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.943767    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:ff0b6c27e2d96afccd73fbbb5b5297a3f60c7f4f1dfd2a877152466697018d71,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pv2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-c87fff755-krm9d_openstack-operators(59abaeff-0c1b-4f9d-bd55-448a1295f603): ErrImagePull: pull QPS exceeded" logger="UnhandledError"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.945528    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d" podUID="59abaeff-0c1b-4f9d-bd55-448a1295f603"
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.946234    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34941af2_d22e_4aaf_b1b6_a620cfc975be.slice/crio-3c6e6782f9c044a1ebcfa9e637c28afe4e41e82fafde142a55f316a71aed3e50 WatchSource:0}: Error finding container 3c6e6782f9c044a1ebcfa9e637c28afe4e41e82fafde142a55f316a71aed3e50: Status 404 returned error can't find the container with id 3c6e6782f9c044a1ebcfa9e637c28afe4e41e82fafde142a55f316a71aed3e50
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.946543    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mbczv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-69797bbcbd-7wxgh_openstack-operators(14b5c974-8804-44d3-9b48-f28f0acab0b6): ErrImagePull: pull QPS exceeded" logger="UnhandledError"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.946907    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q7ckl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-85cd9769bb-rcxtb_openstack-operators(2b9855d9-90b7-4904-9a65-a06a4674d86e): ErrImagePull: pull QPS exceeded" logger="UnhandledError"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.948137    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb" podUID="2b9855d9-90b7-4904-9a65-a06a4674d86e"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.948169    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh" podUID="14b5c974-8804-44d3-9b48-f28f0acab0b6"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.957817    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:2d6d13b3c28e45c6bec980b8808dda8da4723ae87e66d04f53d52c3b3c51612b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x5zjl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5ffb9c6597-84f2h_openstack-operators(34941af2-d22e-4aaf-b1b6-a620cfc975be): ErrImagePull: pull QPS exceeded" logger="UnhandledError"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.959609    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h" podUID="34941af2-d22e-4aaf-b1b6-a620cfc975be"
Jan 22 10:01:35 crc kubenswrapper[4824]: W0122 10:01:35.962373    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4d0a1a7_c43d_461e_998a_c3bfb7ae0a93.slice/crio-328adab2c5ff36ca348ce1da931b0cbbb8e60e6ecfe52a9adb417eec18651698 WatchSource:0}: Error finding container 328adab2c5ff36ca348ce1da931b0cbbb8e60e6ecfe52a9adb417eec18651698: Status 404 returned error can't find the container with id 328adab2c5ff36ca348ce1da931b0cbbb8e60e6ecfe52a9adb417eec18651698
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.963516    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wtmmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-mhlsp_openstack-operators(18d7f488-bb97-4a72-8fbc-eb5db501db16): ErrImagePull: pull QPS exceeded" logger="UnhandledError"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.968533    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp" podUID="18d7f488-bb97-4a72-8fbc-eb5db501db16"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.978415    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8rn6m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-78c6999f6f-zwfpg_openstack-operators(f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93): ErrImagePull: pull QPS exceeded" logger="UnhandledError"
Jan 22 10:01:35 crc kubenswrapper[4824]: E0122 10:01:35.980433    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg" podUID="f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.397231    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.397319    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.397678    4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.397742    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:38.397725026 +0000 UTC m=+1131.301202019 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "webhook-server-cert" not found
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.398160    4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.398241    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:38.398228451 +0000 UTC m=+1131.301705444 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "metrics-server-cert" not found
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.714034    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc" event={"ID":"aa122c78-b0a4-4bd5-9844-7a44b150dac4","Type":"ContainerStarted","Data":"f23e146657938b961c8e2212ec5852c22d771338e3995d6ebfc890f8c93f2efc"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.716097    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt" event={"ID":"cb110aec-f41e-4bae-b0de-9c076a41a603","Type":"ContainerStarted","Data":"15e86aed2a025b1875e80d48130a44bf01a16d42c8343de55532af3c26fe56a5"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.718865    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp" event={"ID":"18d7f488-bb97-4a72-8fbc-eb5db501db16","Type":"ContainerStarted","Data":"aeee3a9918baeae654caaa07c250a7d43a4710addaa4b4d822261198244b9f81"}
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.721988    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp" podUID="18d7f488-bb97-4a72-8fbc-eb5db501db16"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.732393    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5" event={"ID":"2a362e5c-b09f-4c6c-8639-6c7107e049b0","Type":"ContainerStarted","Data":"f66108563d33607d8976ccc9bb81a4b5e0f734cef9e8412a7766d52da1be00ad"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.738975    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg" event={"ID":"f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93","Type":"ContainerStarted","Data":"328adab2c5ff36ca348ce1da931b0cbbb8e60e6ecfe52a9adb417eec18651698"}
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.741230    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg" podUID="f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.753503    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h" event={"ID":"34941af2-d22e-4aaf-b1b6-a620cfc975be","Type":"ContainerStarted","Data":"3c6e6782f9c044a1ebcfa9e637c28afe4e41e82fafde142a55f316a71aed3e50"}
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.756845    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:2d6d13b3c28e45c6bec980b8808dda8da4723ae87e66d04f53d52c3b3c51612b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h" podUID="34941af2-d22e-4aaf-b1b6-a620cfc975be"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.760903    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh" event={"ID":"2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9","Type":"ContainerStarted","Data":"94d6eef35828f6fa1b5fcf0753dc19583fa95f098c0077ca39e2a8123b9d4d49"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.762605    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb" event={"ID":"2b9855d9-90b7-4904-9a65-a06a4674d86e","Type":"ContainerStarted","Data":"4ef51980a141924c38c1e915e1b024c7ff98c5f20fdc2c44315f774a8af7f30a"}
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.764330    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb" podUID="2b9855d9-90b7-4904-9a65-a06a4674d86e"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.764973    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p" event={"ID":"3370a46a-f9bf-4ce2-977b-6e910b3a4448","Type":"ContainerStarted","Data":"df757cf1ad05b5ab74f7786d8b44354a19fafb99b1727e899b5712ad510daa6b"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.766687    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d" event={"ID":"59abaeff-0c1b-4f9d-bd55-448a1295f603","Type":"ContainerStarted","Data":"2ddddd679203698cb9fae800c0cff0851a1f2a23eac31cdd3905d1dcf672684e"}
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.770425    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:ff0b6c27e2d96afccd73fbbb5b5297a3f60c7f4f1dfd2a877152466697018d71\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d" podUID="59abaeff-0c1b-4f9d-bd55-448a1295f603"
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.778981    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh" event={"ID":"14b5c974-8804-44d3-9b48-f28f0acab0b6","Type":"ContainerStarted","Data":"edf8c4a65fe4df94481aec815cdc797b16e230e3148aae64672c016b84cce240"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.781025    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6" event={"ID":"651ec3cd-3884-41c6-ae9d-1fb08b764a32","Type":"ContainerStarted","Data":"4c6532eac33f970c230cb31afe56a634506f20c58be3cbdfaac3d921197cede9"}
Jan 22 10:01:36 crc kubenswrapper[4824]: I0122 10:01:36.781871    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr" event={"ID":"0cf0de8e-51d5-4f14-a63b-08acab6d68a7","Type":"ContainerStarted","Data":"bb2cc9d975f58ce1d5ef978c01c4dadc7cf19912132b3fba2f66e012b58ddf25"}
Jan 22 10:01:36 crc kubenswrapper[4824]: E0122 10:01:36.784251    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh" podUID="14b5c974-8804-44d3-9b48-f28f0acab0b6"
Jan 22 10:01:37 crc kubenswrapper[4824]: I0122 10:01:37.722955    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.723093    4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.723153    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert podName:40210948-0e80-4f69-9d75-ab44c0d38285 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:41.723138697 +0000 UTC m=+1134.626615690 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert") pod "infra-operator-controller-manager-54ccf4f85d-gbhz2" (UID: "40210948-0e80-4f69-9d75-ab44c0d38285") : secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.792596    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:ff0b6c27e2d96afccd73fbbb5b5297a3f60c7f4f1dfd2a877152466697018d71\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d" podUID="59abaeff-0c1b-4f9d-bd55-448a1295f603"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.792658    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb" podUID="2b9855d9-90b7-4904-9a65-a06a4674d86e"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.793070    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:2d6d13b3c28e45c6bec980b8808dda8da4723ae87e66d04f53d52c3b3c51612b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h" podUID="34941af2-d22e-4aaf-b1b6-a620cfc975be"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.793206    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh" podUID="14b5c974-8804-44d3-9b48-f28f0acab0b6"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.793293    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp" podUID="18d7f488-bb97-4a72-8fbc-eb5db501db16"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.794514    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg" podUID="f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93"
Jan 22 10:01:37 crc kubenswrapper[4824]: I0122 10:01:37.925143    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.925298    4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:37 crc kubenswrapper[4824]: E0122 10:01:37.925371    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert podName:d5b9d0a8-b7a8-4138-a306-4599159f33e1 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:41.92533919 +0000 UTC m=+1134.828816183 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" (UID: "d5b9d0a8-b7a8-4138-a306-4599159f33e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:38 crc kubenswrapper[4824]: I0122 10:01:38.437406    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:38 crc kubenswrapper[4824]: I0122 10:01:38.437492    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:38 crc kubenswrapper[4824]: E0122 10:01:38.437600    4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found
Jan 22 10:01:38 crc kubenswrapper[4824]: E0122 10:01:38.437667    4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found
Jan 22 10:01:38 crc kubenswrapper[4824]: E0122 10:01:38.437672    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:42.437653205 +0000 UTC m=+1135.341130198 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "metrics-server-cert" not found
Jan 22 10:01:38 crc kubenswrapper[4824]: E0122 10:01:38.437727    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:42.437713797 +0000 UTC m=+1135.341190790 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "webhook-server-cert" not found
Jan 22 10:01:41 crc kubenswrapper[4824]: I0122 10:01:41.786253    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:41 crc kubenswrapper[4824]: E0122 10:01:41.786573    4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:41 crc kubenswrapper[4824]: E0122 10:01:41.788334    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert podName:40210948-0e80-4f69-9d75-ab44c0d38285 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:49.788312647 +0000 UTC m=+1142.691789640 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert") pod "infra-operator-controller-manager-54ccf4f85d-gbhz2" (UID: "40210948-0e80-4f69-9d75-ab44c0d38285") : secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:41 crc kubenswrapper[4824]: E0122 10:01:41.991530    4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:41 crc kubenswrapper[4824]: E0122 10:01:41.991944    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert podName:d5b9d0a8-b7a8-4138-a306-4599159f33e1 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:49.991917269 +0000 UTC m=+1142.895394302 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert") pod "openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" (UID: "d5b9d0a8-b7a8-4138-a306-4599159f33e1") : secret "openstack-baremetal-operator-webhook-server-cert" not found
Jan 22 10:01:41 crc kubenswrapper[4824]: I0122 10:01:41.991344    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:42 crc kubenswrapper[4824]: E0122 10:01:42.501583    4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found
Jan 22 10:01:42 crc kubenswrapper[4824]: E0122 10:01:42.501673    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:50.501651802 +0000 UTC m=+1143.405128845 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "metrics-server-cert" not found
Jan 22 10:01:42 crc kubenswrapper[4824]: I0122 10:01:42.500278    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:42 crc kubenswrapper[4824]: I0122 10:01:42.502341    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:42 crc kubenswrapper[4824]: E0122 10:01:42.502798    4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found
Jan 22 10:01:42 crc kubenswrapper[4824]: E0122 10:01:42.502860    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs podName:12e59f22-9370-4b23-8d8a-5c24ba925541 nodeName:}" failed. No retries permitted until 2026-01-22 10:01:50.502842006 +0000 UTC m=+1143.406319069 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs") pod "openstack-operator-controller-manager-788c8b99b5-fmxvw" (UID: "12e59f22-9370-4b23-8d8a-5c24ba925541") : secret "webhook-server-cert" not found
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.070826    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.071343    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gxnjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5d646b7d76-vgkz5_openstack-operators(2a362e5c-b09f-4c6c-8639-6c7107e049b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.072570    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5" podUID="2a362e5c-b09f-4c6c-8639-6c7107e049b0"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.793781    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.793981    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t5d8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-b45d7bf98-8prhs_openstack-operators(905ddc8e-1123-4654-92a3-a50493b92a78): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.795109    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs" podUID="905ddc8e-1123-4654-92a3-a50493b92a78"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.880911    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece\\\"\"" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs" podUID="905ddc8e-1123-4654-92a3-a50493b92a78"
Jan 22 10:01:48 crc kubenswrapper[4824]: E0122 10:01:48.882541    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5" podUID="2a362e5c-b09f-4c6c-8639-6c7107e049b0"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.288548    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.288775    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hkksm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-78fdd796fd-rt6zn_openstack-operators(b53e7faf-8d7c-4f40-82c7-e6aa64508488): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.289938    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn" podUID="b53e7faf-8d7c-4f40-82c7-e6aa64508488"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.762142    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:e5e017be64edd679623ea1b7e6a1ae780fdcee4ef79be989b93d8c1d082da15b"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.762305    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:e5e017be64edd679623ea1b7e6a1ae780fdcee4ef79be989b93d8c1d082da15b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mpt6d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-59dd8b7cbf-nbrrg_openstack-operators(88e0e1fa-e0ae-43e1-b653-158ea72244b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.763475    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg" podUID="88e0e1fa-e0ae-43e1-b653-158ea72244b1"
Jan 22 10:01:49 crc kubenswrapper[4824]: I0122 10:01:49.809609    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.809801    4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.809961    4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert podName:40210948-0e80-4f69-9d75-ab44c0d38285 nodeName:}" failed. No retries permitted until 2026-01-22 10:02:05.809943242 +0000 UTC m=+1158.713420235 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert") pod "infra-operator-controller-manager-54ccf4f85d-gbhz2" (UID: "40210948-0e80-4f69-9d75-ab44c0d38285") : secret "infra-operator-webhook-server-cert" not found
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.884061    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337\\\"\"" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn" podUID="b53e7faf-8d7c-4f40-82c7-e6aa64508488"
Jan 22 10:01:49 crc kubenswrapper[4824]: E0122 10:01:49.884458    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:e5e017be64edd679623ea1b7e6a1ae780fdcee4ef79be989b93d8c1d082da15b\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg" podUID="88e0e1fa-e0ae-43e1-b653-158ea72244b1"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.012772    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.021989    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5b9d0a8-b7a8-4138-a306-4599159f33e1-cert\") pod \"openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f\" (UID: \"d5b9d0a8-b7a8-4138-a306-4599159f33e1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.200316    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-mbmxr"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.209111    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.235064    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.235548    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wp6sj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-594c8c9d5d-f5qrl_openstack-operators(792e3f5e-1236-4fe0-94b5-20e322aae84c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.236854    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl" podUID="792e3f5e-1236-4fe0-94b5-20e322aae84c"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.527645    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.527907    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.532221    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-metrics-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.533043    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/12e59f22-9370-4b23-8d8a-5c24ba925541-webhook-certs\") pod \"openstack-operator-controller-manager-788c8b99b5-fmxvw\" (UID: \"12e59f22-9370-4b23-8d8a-5c24ba925541\") " pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.639262    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-sphf7"
Jan 22 10:01:50 crc kubenswrapper[4824]: I0122 10:01:50.647723    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.735922    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.736119    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fxk9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-547cbdb99f-ch89p_openstack-operators(3370a46a-f9bf-4ce2-977b-6e910b3a4448): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.737305    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p" podUID="3370a46a-f9bf-4ce2-977b-6e910b3a4448"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.892300    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p" podUID="3370a46a-f9bf-4ce2-977b-6e910b3a4448"
Jan 22 10:01:50 crc kubenswrapper[4824]: E0122 10:01:50.892639    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492\\\"\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl" podUID="792e3f5e-1236-4fe0-94b5-20e322aae84c"
Jan 22 10:01:51 crc kubenswrapper[4824]: I0122 10:01:51.451515    4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Jan 22 10:01:52 crc kubenswrapper[4824]: E0122 10:01:52.545439    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf"
Jan 22 10:01:52 crc kubenswrapper[4824]: E0122 10:01:52.545870    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l84rg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-55db956ddc-l8kpr_openstack-operators(0cf0de8e-51d5-4f14-a63b-08acab6d68a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:52 crc kubenswrapper[4824]: E0122 10:01:52.547979    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr" podUID="0cf0de8e-51d5-4f14-a63b-08acab6d68a7"
Jan 22 10:01:52 crc kubenswrapper[4824]: E0122 10:01:52.905723    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr" podUID="0cf0de8e-51d5-4f14-a63b-08acab6d68a7"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.197512    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:b57d65d2a968705b9067192a7cb33bd4a12489db87e1d05de78c076f2062cab4"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.197680    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:b57d65d2a968705b9067192a7cb33bd4a12489db87e1d05de78c076f2062cab4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s8p7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5d8f59fb49-q25f6_openstack-operators(651ec3cd-3884-41c6-ae9d-1fb08b764a32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.198894    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6" podUID="651ec3cd-3884-41c6-ae9d-1fb08b764a32"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.759271    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.759465    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9z78c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b8b6d4659-jkg9n_openstack-operators(1b0c67df-ba69-46b3-ad14-e7287f34fd27): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.761440    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n" podUID="1b0c67df-ba69-46b3-ad14-e7287f34fd27"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.911121    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n" podUID="1b0c67df-ba69-46b3-ad14-e7287f34fd27"
Jan 22 10:01:53 crc kubenswrapper[4824]: E0122 10:01:53.912032    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:b57d65d2a968705b9067192a7cb33bd4a12489db87e1d05de78c076f2062cab4\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6" podUID="651ec3cd-3884-41c6-ae9d-1fb08b764a32"
Jan 22 10:01:55 crc kubenswrapper[4824]: E0122 10:01:55.523914    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831"
Jan 22 10:01:55 crc kubenswrapper[4824]: E0122 10:01:55.524142    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fd5w7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-6b8bc8d87d-hrzwc_openstack-operators(aa122c78-b0a4-4bd5-9844-7a44b150dac4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:01:55 crc kubenswrapper[4824]: E0122 10:01:55.525450    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc" podUID="aa122c78-b0a4-4bd5-9844-7a44b150dac4"
Jan 22 10:01:55 crc kubenswrapper[4824]: E0122 10:01:55.946799    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc" podUID="aa122c78-b0a4-4bd5-9844-7a44b150dac4"
Jan 22 10:01:58 crc kubenswrapper[4824]: I0122 10:01:58.977531    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm" event={"ID":"6507116d-ad87-4be1-8322-340170f575b3","Type":"ContainerStarted","Data":"f3d16c0a82c6d62527c06d2c91107070dfb1e207caa860cd9cdd378eabd07bf5"}
Jan 22 10:01:58 crc kubenswrapper[4824]: I0122 10:01:58.983239    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.018078    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm" podStartSLOduration=6.27565512 podStartE2EDuration="26.018056611s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.093642955 +0000 UTC m=+1127.997119948" lastFinishedPulling="2026-01-22 10:01:54.836044446 +0000 UTC m=+1147.739521439" observedRunningTime="2026-01-22 10:01:59.01556414 +0000 UTC m=+1151.919041133" watchObservedRunningTime="2026-01-22 10:01:59.018056611 +0000 UTC m=+1151.921533604"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.142059    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"]
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.220865    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"]
Jan 22 10:01:59 crc kubenswrapper[4824]: W0122 10:01:59.221710    4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12e59f22_9370_4b23_8d8a_5c24ba925541.slice/crio-7f212a47846b7fdce749ed3abac0b71339f1e423eddb7bea71be519eb7e342f6 WatchSource:0}: Error finding container 7f212a47846b7fdce749ed3abac0b71339f1e423eddb7bea71be519eb7e342f6: Status 404 returned error can't find the container with id 7f212a47846b7fdce749ed3abac0b71339f1e423eddb7bea71be519eb7e342f6
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.984767    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt" event={"ID":"cb110aec-f41e-4bae-b0de-9c076a41a603","Type":"ContainerStarted","Data":"c60f080c82c5e36c01a7040b88894fde4908cd9f454d87b59b2bb58ae0841910"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.985692    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.986591    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh" event={"ID":"2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9","Type":"ContainerStarted","Data":"176cff0a8e458861786628260266a9c478c07e5891f4e6a4e05b6970643211e6"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.986943    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.988128    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb" event={"ID":"2b9855d9-90b7-4904-9a65-a06a4674d86e","Type":"ContainerStarted","Data":"9aaf0e1882b17fef53909334ec0da18493a1425c9e634ccae927e41551dd1eba"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.988498    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.989781    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg" event={"ID":"f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93","Type":"ContainerStarted","Data":"52817f06ff2fd62e4797f4b003038840d28139f891f68416c1651e7d263b6764"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.990206    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.991147    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" event={"ID":"d5b9d0a8-b7a8-4138-a306-4599159f33e1","Type":"ContainerStarted","Data":"b15fe9bc09d7d0b5de6fbf0e9fe42c18bcfc2254a749000523fd993577519c44"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.992544    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh" event={"ID":"14b5c974-8804-44d3-9b48-f28f0acab0b6","Type":"ContainerStarted","Data":"27a4f074a07ee3d4a5ce273ed5834e9d66fc139410ad076c6e997a39d8c43eb8"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.992994    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.994314    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp" event={"ID":"18d7f488-bb97-4a72-8fbc-eb5db501db16","Type":"ContainerStarted","Data":"79f791e96aeb1e0475f58db233cf3b36c3764070846218400f6ab4b07bbc67fa"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.995865    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt" event={"ID":"d29a7a0c-344a-449a-bfa0-457e860f9030","Type":"ContainerStarted","Data":"7e0bf2b938e400485eb1f7498dd10b43059d4513682cb1be0bdbd56240e1cdeb"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.996232    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.997267    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h" event={"ID":"34941af2-d22e-4aaf-b1b6-a620cfc975be","Type":"ContainerStarted","Data":"b2b3218561196d87ac0f29438a94c760ffeee2f2410879b5cbb8ef5abe855125"}
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.997623    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:01:59 crc kubenswrapper[4824]: I0122 10:01:59.999789    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d" event={"ID":"59abaeff-0c1b-4f9d-bd55-448a1295f603","Type":"ContainerStarted","Data":"06e06561a00069701e82e0ab3b577f79904047796625f898ef55496bed3efeac"}
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.000427    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.002791    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw" event={"ID":"12e59f22-9370-4b23-8d8a-5c24ba925541","Type":"ContainerStarted","Data":"553c308ff7dc9feeab4bba5aa1cc751b489b27a2f276970ff80e1cebeb6bf8d7"}
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.002830    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw" event={"ID":"12e59f22-9370-4b23-8d8a-5c24ba925541","Type":"ContainerStarted","Data":"7f212a47846b7fdce749ed3abac0b71339f1e423eddb7bea71be519eb7e342f6"}
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.002848    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.005100    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt" podStartSLOduration=7.3011359670000004 podStartE2EDuration="27.00507804s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.839668594 +0000 UTC m=+1128.743145587" lastFinishedPulling="2026-01-22 10:01:55.543610667 +0000 UTC m=+1148.447087660" observedRunningTime="2026-01-22 10:02:00.000165029 +0000 UTC m=+1152.903642022" watchObservedRunningTime="2026-01-22 10:02:00.00507804 +0000 UTC m=+1152.908555033"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.046423    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh" podStartSLOduration=3.3452098550000002 podStartE2EDuration="26.046403959s" podCreationTimestamp="2026-01-22 10:01:34 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.946453272 +0000 UTC m=+1128.849930265" lastFinishedPulling="2026-01-22 10:01:58.647647376 +0000 UTC m=+1151.551124369" observedRunningTime="2026-01-22 10:02:00.026306916 +0000 UTC m=+1152.929783919" watchObservedRunningTime="2026-01-22 10:02:00.046403959 +0000 UTC m=+1152.949880962"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.136236    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h" podStartSLOduration=3.446708364 podStartE2EDuration="26.136215383s" podCreationTimestamp="2026-01-22 10:01:34 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.95756103 +0000 UTC m=+1128.861038023" lastFinishedPulling="2026-01-22 10:01:58.647068049 +0000 UTC m=+1151.550545042" observedRunningTime="2026-01-22 10:02:00.055198101 +0000 UTC m=+1152.958675094" watchObservedRunningTime="2026-01-22 10:02:00.136215383 +0000 UTC m=+1153.039692386"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.140764    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg" podStartSLOduration=4.491942536 podStartE2EDuration="27.140746003s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.978269071 +0000 UTC m=+1128.881746064" lastFinishedPulling="2026-01-22 10:01:58.627072538 +0000 UTC m=+1151.530549531" observedRunningTime="2026-01-22 10:02:00.134413072 +0000 UTC m=+1153.037890075" watchObservedRunningTime="2026-01-22 10:02:00.140746003 +0000 UTC m=+1153.044222996"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.280560    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt" podStartSLOduration=7.420847444 podStartE2EDuration="27.280544584s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:34.977040096 +0000 UTC m=+1127.880517089" lastFinishedPulling="2026-01-22 10:01:54.836737236 +0000 UTC m=+1147.740214229" observedRunningTime="2026-01-22 10:02:00.222565529 +0000 UTC m=+1153.126042522" watchObservedRunningTime="2026-01-22 10:02:00.280544584 +0000 UTC m=+1153.184021577"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.282677    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh" podStartSLOduration=7.558430122 podStartE2EDuration="27.282670865s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.818098638 +0000 UTC m=+1128.721575631" lastFinishedPulling="2026-01-22 10:01:55.542339381 +0000 UTC m=+1148.445816374" observedRunningTime="2026-01-22 10:02:00.278483735 +0000 UTC m=+1153.181960728" watchObservedRunningTime="2026-01-22 10:02:00.282670865 +0000 UTC m=+1153.186147858"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.298940    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d" podStartSLOduration=4.611350175 podStartE2EDuration="27.298918389s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.943624972 +0000 UTC m=+1128.847101965" lastFinishedPulling="2026-01-22 10:01:58.631193186 +0000 UTC m=+1151.534670179" observedRunningTime="2026-01-22 10:02:00.294989096 +0000 UTC m=+1153.198466089" watchObservedRunningTime="2026-01-22 10:02:00.298918389 +0000 UTC m=+1153.202395392"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.311045    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb" podStartSLOduration=3.6278806660000003 podStartE2EDuration="26.311027114s" podCreationTimestamp="2026-01-22 10:01:34 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.946847254 +0000 UTC m=+1128.850324247" lastFinishedPulling="2026-01-22 10:01:58.629993702 +0000 UTC m=+1151.533470695" observedRunningTime="2026-01-22 10:02:00.309551362 +0000 UTC m=+1153.213028365" watchObservedRunningTime="2026-01-22 10:02:00.311027114 +0000 UTC m=+1153.214504107"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.325265    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mhlsp" podStartSLOduration=3.5306115890000003 podStartE2EDuration="26.32524407s" podCreationTimestamp="2026-01-22 10:01:34 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.963408146 +0000 UTC m=+1128.866885139" lastFinishedPulling="2026-01-22 10:01:58.758040627 +0000 UTC m=+1151.661517620" observedRunningTime="2026-01-22 10:02:00.323240233 +0000 UTC m=+1153.226717226" watchObservedRunningTime="2026-01-22 10:02:00.32524407 +0000 UTC m=+1153.228721063"
Jan 22 10:02:00 crc kubenswrapper[4824]: I0122 10:02:00.369553    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw" podStartSLOduration=26.369532135 podStartE2EDuration="26.369532135s" podCreationTimestamp="2026-01-22 10:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-22 10:02:00.362831023 +0000 UTC m=+1153.266308026" watchObservedRunningTime="2026-01-22 10:02:00.369532135 +0000 UTC m=+1153.273009118"
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.040875    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn" event={"ID":"b53e7faf-8d7c-4f40-82c7-e6aa64508488","Type":"ContainerStarted","Data":"58837f87f1d76575be8daad08f0d9a56cd87bf1412e48e32fa1af53b1b7c2ad5"}
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.041467    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.042858    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl" event={"ID":"792e3f5e-1236-4fe0-94b5-20e322aae84c","Type":"ContainerStarted","Data":"60c778cc36af115bbe9d2e22efb7bad46e80e8967917fafa3f42646657fac020"}
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.043997    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.119968    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn" podStartSLOduration=3.321090849 podStartE2EDuration="30.119946199s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.595776331 +0000 UTC m=+1128.499253324" lastFinishedPulling="2026-01-22 10:02:02.394631681 +0000 UTC m=+1155.298108674" observedRunningTime="2026-01-22 10:02:03.105815735 +0000 UTC m=+1156.009292728" watchObservedRunningTime="2026-01-22 10:02:03.119946199 +0000 UTC m=+1156.023423192"
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.969499    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-vshwt"
Jan 22 10:02:03 crc kubenswrapper[4824]: I0122 10:02:03.988343    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl" podStartSLOduration=4.301073216 podStartE2EDuration="30.98832179s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.63253771 +0000 UTC m=+1128.536014703" lastFinishedPulling="2026-01-22 10:02:02.319786284 +0000 UTC m=+1155.223263277" observedRunningTime="2026-01-22 10:02:03.122733028 +0000 UTC m=+1156.026210031" watchObservedRunningTime="2026-01-22 10:02:03.98832179 +0000 UTC m=+1156.891798783"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.057099    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg" event={"ID":"88e0e1fa-e0ae-43e1-b653-158ea72244b1","Type":"ContainerStarted","Data":"b903dfe71c5cc81903a242c9a2ef1230ebf60ff40c8b63dda366c2d914e18d7b"}
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.058117    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.217523    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-k2hwm"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.239625    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg" podStartSLOduration=3.461439595 podStartE2EDuration="31.239605404s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.417491781 +0000 UTC m=+1128.320968774" lastFinishedPulling="2026-01-22 10:02:03.19565759 +0000 UTC m=+1156.099134583" observedRunningTime="2026-01-22 10:02:04.074816629 +0000 UTC m=+1156.978293622" watchObservedRunningTime="2026-01-22 10:02:04.239605404 +0000 UTC m=+1157.143082397"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.368708    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-69d6c9f5b8-npvrh"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.418245    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-zwfpg"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.451977    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-krm9d"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.861321    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-rcxtb"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.943813    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-7wxgh"
Jan 22 10:02:04 crc kubenswrapper[4824]: I0122 10:02:04.955773    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5ffb9c6597-84f2h"
Jan 22 10:02:05 crc kubenswrapper[4824]: I0122 10:02:05.588620    4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt" podUID="cb110aec-f41e-4bae-b0de-9c076a41a603" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)"
Jan 22 10:02:05 crc kubenswrapper[4824]: I0122 10:02:05.839649    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:02:05 crc kubenswrapper[4824]: I0122 10:02:05.849679    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/40210948-0e80-4f69-9d75-ab44c0d38285-cert\") pod \"infra-operator-controller-manager-54ccf4f85d-gbhz2\" (UID: \"40210948-0e80-4f69-9d75-ab44c0d38285\") " pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:02:06 crc kubenswrapper[4824]: I0122 10:02:06.038760    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jhzqr"
Jan 22 10:02:06 crc kubenswrapper[4824]: I0122 10:02:06.047221    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:02:08 crc kubenswrapper[4824]: I0122 10:02:08.436416    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"]
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.179490    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6" event={"ID":"651ec3cd-3884-41c6-ae9d-1fb08b764a32","Type":"ContainerStarted","Data":"bbcf5abe5a90db99a1ccd0bcde07bfc043086c7c5b3df8db02e1d830c19319dd"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.180003    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.190278    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr" event={"ID":"0cf0de8e-51d5-4f14-a63b-08acab6d68a7","Type":"ContainerStarted","Data":"eec746994d0ba1ec66a19ce96e2cd8b0365963fc302bd3e7ccccb3f7f117ef03"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.190533    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.200558    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6" podStartSLOduration=4.118901266 podStartE2EDuration="36.200538058s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.920766199 +0000 UTC m=+1128.824243192" lastFinishedPulling="2026-01-22 10:02:08.002402991 +0000 UTC m=+1160.905879984" observedRunningTime="2026-01-22 10:02:09.193384244 +0000 UTC m=+1162.096861237" watchObservedRunningTime="2026-01-22 10:02:09.200538058 +0000 UTC m=+1162.104015051"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.202809    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs" event={"ID":"905ddc8e-1123-4654-92a3-a50493b92a78","Type":"ContainerStarted","Data":"63cd42ad38299e941bc8fa86157598b185c97d9a2373a09b7a15b88fa7bf828d"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.203365    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.204741    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5" event={"ID":"2a362e5c-b09f-4c6c-8639-6c7107e049b0","Type":"ContainerStarted","Data":"a38646f7d752779cce976a904b8889bb209c41945a753985913efc97739e5310"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.204946    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.210320    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p" event={"ID":"3370a46a-f9bf-4ce2-977b-6e910b3a4448","Type":"ContainerStarted","Data":"676a7b1fb0fb879c6a5656c816a84cd4512b4084940c410a7bab42b24a1d8b59"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.210595    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.215680    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" event={"ID":"d5b9d0a8-b7a8-4138-a306-4599159f33e1","Type":"ContainerStarted","Data":"c65e15554f58a20e319b5ccf13267df4dae7efcf5c9250a2b94a0893bb84514f"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.216235    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.217475    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2" event={"ID":"40210948-0e80-4f69-9d75-ab44c0d38285","Type":"ContainerStarted","Data":"d7725a569980f0d0ff4d29f8c5eb411cbc932a10da9667a27b6120915e3195e9"}
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.252313    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr" podStartSLOduration=4.112910585 podStartE2EDuration="36.252293165s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.863842444 +0000 UTC m=+1128.767319437" lastFinishedPulling="2026-01-22 10:02:08.003225024 +0000 UTC m=+1160.906702017" observedRunningTime="2026-01-22 10:02:09.249888127 +0000 UTC m=+1162.153365120" watchObservedRunningTime="2026-01-22 10:02:09.252293165 +0000 UTC m=+1162.155770158"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.288081    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs" podStartSLOduration=3.91902959 podStartE2EDuration="36.288059967s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.637568394 +0000 UTC m=+1128.541045387" lastFinishedPulling="2026-01-22 10:02:08.006598771 +0000 UTC m=+1160.910075764" observedRunningTime="2026-01-22 10:02:09.267313184 +0000 UTC m=+1162.170790167" watchObservedRunningTime="2026-01-22 10:02:09.288059967 +0000 UTC m=+1162.191536960"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.289636    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5" podStartSLOduration=4.153675018 podStartE2EDuration="36.289626941s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.870508454 +0000 UTC m=+1128.773985447" lastFinishedPulling="2026-01-22 10:02:08.006460377 +0000 UTC m=+1160.909937370" observedRunningTime="2026-01-22 10:02:09.285077491 +0000 UTC m=+1162.188554484" watchObservedRunningTime="2026-01-22 10:02:09.289626941 +0000 UTC m=+1162.193103944"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.302711    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p" podStartSLOduration=3.770285803 podStartE2EDuration="36.302695224s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.878731379 +0000 UTC m=+1128.782208372" lastFinishedPulling="2026-01-22 10:02:08.4111408 +0000 UTC m=+1161.314617793" observedRunningTime="2026-01-22 10:02:09.300272355 +0000 UTC m=+1162.203749348" watchObservedRunningTime="2026-01-22 10:02:09.302695224 +0000 UTC m=+1162.206172217"
Jan 22 10:02:09 crc kubenswrapper[4824]: I0122 10:02:09.331989    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f" podStartSLOduration=27.357096081 podStartE2EDuration="36.3319704s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:59.146250271 +0000 UTC m=+1152.049727264" lastFinishedPulling="2026-01-22 10:02:08.12112459 +0000 UTC m=+1161.024601583" observedRunningTime="2026-01-22 10:02:09.324842157 +0000 UTC m=+1162.228319150" watchObservedRunningTime="2026-01-22 10:02:09.3319704 +0000 UTC m=+1162.235447393"
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.225320    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc" event={"ID":"aa122c78-b0a4-4bd5-9844-7a44b150dac4","Type":"ContainerStarted","Data":"f25b6af22b9d130fe2825bc6143723b26c309ca37a376494631985227b1f2e5f"}
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.225504    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.229273    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n" event={"ID":"1b0c67df-ba69-46b3-ad14-e7287f34fd27","Type":"ContainerStarted","Data":"ff63e3370c0ba04a8eba124c85d0a0f173a6f2cd71c7ec11def0e9eafc8dc59b"}
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.229715    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.253812    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc" podStartSLOduration=4.590950172 podStartE2EDuration="37.253797038s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.840888008 +0000 UTC m=+1128.744365001" lastFinishedPulling="2026-01-22 10:02:08.503734874 +0000 UTC m=+1161.407211867" observedRunningTime="2026-01-22 10:02:10.251005009 +0000 UTC m=+1163.154482002" watchObservedRunningTime="2026-01-22 10:02:10.253797038 +0000 UTC m=+1163.157274031"
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.273256    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n" podStartSLOduration=3.587172835 podStartE2EDuration="37.273233133s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:01:35.416942555 +0000 UTC m=+1128.320419558" lastFinishedPulling="2026-01-22 10:02:09.103002863 +0000 UTC m=+1162.006479856" observedRunningTime="2026-01-22 10:02:10.266837301 +0000 UTC m=+1163.170314284" watchObservedRunningTime="2026-01-22 10:02:10.273233133 +0000 UTC m=+1163.176710126"
Jan 22 10:02:10 crc kubenswrapper[4824]: I0122 10:02:10.656558    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-788c8b99b5-fmxvw"
Jan 22 10:02:13 crc kubenswrapper[4824]: I0122 10:02:13.980159    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59dd8b7cbf-nbrrg"
Jan 22 10:02:13 crc kubenswrapper[4824]: I0122 10:02:13.999822    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-8prhs"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.068207    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-rt6zn"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.106411    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-f5qrl"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.396365    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-jkg9n"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.492447    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5d8f59fb49-q25f6"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.520861    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-hrzwc"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.547498    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-wm2vt"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.687886    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-vgkz5"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.739899    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-l8kpr"
Jan 22 10:02:14 crc kubenswrapper[4824]: I0122 10:02:14.760241    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-ch89p"
Jan 22 10:02:20 crc kubenswrapper[4824]: I0122 10:02:20.215658    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f"
Jan 22 10:02:22 crc kubenswrapper[4824]: E0122 10:02:22.359369    4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:2eac1b9dadaddf4734f35e3dd1996dca960e97d2f304cbd48254b900a840a84a"
Jan 22 10:02:22 crc kubenswrapper[4824]: E0122 10:02:22.359562    4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:2eac1b9dadaddf4734f35e3dd1996dca960e97d2f304cbd48254b900a840a84a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {}  BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rwch5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-54ccf4f85d-gbhz2_openstack-operators(40210948-0e80-4f69-9d75-ab44c0d38285): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError"
Jan 22 10:02:22 crc kubenswrapper[4824]: E0122 10:02:22.360934    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2" podUID="40210948-0e80-4f69-9d75-ab44c0d38285"
Jan 22 10:02:22 crc kubenswrapper[4824]: E0122 10:02:22.400429    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:2eac1b9dadaddf4734f35e3dd1996dca960e97d2f304cbd48254b900a840a84a\\\"\"" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2" podUID="40210948-0e80-4f69-9d75-ab44c0d38285"
Jan 22 10:02:30 crc kubenswrapper[4824]: I0122 10:02:30.332907    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:02:30 crc kubenswrapper[4824]: I0122 10:02:30.333470    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:02:35 crc kubenswrapper[4824]: I0122 10:02:35.487893    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2" event={"ID":"40210948-0e80-4f69-9d75-ab44c0d38285","Type":"ContainerStarted","Data":"4fe79edb051254f8dd6cf5001ecd27776f71fc19f10d9a6b5bc44950d51f905b"}
Jan 22 10:02:35 crc kubenswrapper[4824]: I0122 10:02:35.488645    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:02:35 crc kubenswrapper[4824]: I0122 10:02:35.506764    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2" podStartSLOduration=36.102589341 podStartE2EDuration="1m2.506749451s" podCreationTimestamp="2026-01-22 10:01:33 +0000 UTC" firstStartedPulling="2026-01-22 10:02:08.431015438 +0000 UTC m=+1161.334492431" lastFinishedPulling="2026-01-22 10:02:34.835175548 +0000 UTC m=+1187.738652541" observedRunningTime="2026-01-22 10:02:35.504064654 +0000 UTC m=+1188.407541647" watchObservedRunningTime="2026-01-22 10:02:35.506749451 +0000 UTC m=+1188.410226444"
Jan 22 10:02:46 crc kubenswrapper[4824]: I0122 10:02:46.054973    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-54ccf4f85d-gbhz2"
Jan 22 10:03:00 crc kubenswrapper[4824]: I0122 10:03:00.333927    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:03:00 crc kubenswrapper[4824]: I0122 10:03:00.334994    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.237744    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nnxxz/must-gather-t4pqm"]
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.239733    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.242036    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nnxxz"/"openshift-service-ca.crt"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.242119    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nnxxz"/"kube-root-ca.crt"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.242051    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-nnxxz"/"default-dockercfg-8k9z4"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.254433    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nnxxz/must-gather-t4pqm"]
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.349180    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzwj2\" (UniqueName: \"kubernetes.io/projected/39587411-6600-4d7d-9c5a-dc98e8b5c107-kube-api-access-kzwj2\") pod \"must-gather-t4pqm\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") " pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.349384    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/39587411-6600-4d7d-9c5a-dc98e8b5c107-must-gather-output\") pod \"must-gather-t4pqm\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") " pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.450840    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/39587411-6600-4d7d-9c5a-dc98e8b5c107-must-gather-output\") pod \"must-gather-t4pqm\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") " pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.450905    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzwj2\" (UniqueName: \"kubernetes.io/projected/39587411-6600-4d7d-9c5a-dc98e8b5c107-kube-api-access-kzwj2\") pod \"must-gather-t4pqm\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") " pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.451382    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/39587411-6600-4d7d-9c5a-dc98e8b5c107-must-gather-output\") pod \"must-gather-t4pqm\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") " pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.476984    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzwj2\" (UniqueName: \"kubernetes.io/projected/39587411-6600-4d7d-9c5a-dc98e8b5c107-kube-api-access-kzwj2\") pod \"must-gather-t4pqm\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") " pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.555656    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:03:26 crc kubenswrapper[4824]: I0122 10:03:26.965035    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nnxxz/must-gather-t4pqm"]
Jan 22 10:03:27 crc kubenswrapper[4824]: I0122 10:03:27.859997    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nnxxz/must-gather-t4pqm" event={"ID":"39587411-6600-4d7d-9c5a-dc98e8b5c107","Type":"ContainerStarted","Data":"d78be1b14e2b4b9d43e7e4c439aea15b96ec0a6a3a76f8b733204b8c7618877e"}
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.333264    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.333335    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.333407    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.334053    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"142d82198208020ed04e19461fd1e7bda1b5d653c554ec42bffda74db5a4d02d"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.334110    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://142d82198208020ed04e19461fd1e7bda1b5d653c554ec42bffda74db5a4d02d" gracePeriod=600
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.929189    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="142d82198208020ed04e19461fd1e7bda1b5d653c554ec42bffda74db5a4d02d" exitCode=0
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.929234    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"142d82198208020ed04e19461fd1e7bda1b5d653c554ec42bffda74db5a4d02d"}
Jan 22 10:03:30 crc kubenswrapper[4824]: I0122 10:03:30.929272    4824 scope.go:117] "RemoveContainer" containerID="830c35e94e49d8ce59cb2a03a98095eef4c11c2c5bd0a68e880b21c7e9b8661a"
Jan 22 10:03:36 crc kubenswrapper[4824]: I0122 10:03:36.968781    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"363a321edd4efcce621b593f7e6704f72529b44f15f265b34c10f2ff6d4ad874"}
Jan 22 10:03:36 crc kubenswrapper[4824]: I0122 10:03:36.970557    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nnxxz/must-gather-t4pqm" event={"ID":"39587411-6600-4d7d-9c5a-dc98e8b5c107","Type":"ContainerStarted","Data":"f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3"}
Jan 22 10:03:36 crc kubenswrapper[4824]: I0122 10:03:36.970667    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nnxxz/must-gather-t4pqm" event={"ID":"39587411-6600-4d7d-9c5a-dc98e8b5c107","Type":"ContainerStarted","Data":"c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"}
Jan 22 10:03:37 crc kubenswrapper[4824]: I0122 10:03:37.010622    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nnxxz/must-gather-t4pqm" podStartSLOduration=1.964069023 podStartE2EDuration="11.01060493s" podCreationTimestamp="2026-01-22 10:03:26 +0000 UTC" firstStartedPulling="2026-01-22 10:03:26.973486638 +0000 UTC m=+1239.876963631" lastFinishedPulling="2026-01-22 10:03:36.020022545 +0000 UTC m=+1248.923499538" observedRunningTime="2026-01-22 10:03:37.004343784 +0000 UTC m=+1249.907820777" watchObservedRunningTime="2026-01-22 10:03:37.01060493 +0000 UTC m=+1249.914081913"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.234174    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59dd8b7cbf-nbrrg_88e0e1fa-e0ae-43e1-b653-158ea72244b1/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.256159    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-vshwt_d29a7a0c-344a-449a-bfa0-457e860f9030/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.272416    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-8prhs_905ddc8e-1123-4654-92a3-a50493b92a78/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.285601    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/extract/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.296748    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/util/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.308410    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/pull/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.327543    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-rt6zn_b53e7faf-8d7c-4f40-82c7-e6aa64508488/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.342005    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-f5qrl_792e3f5e-1236-4fe0-94b5-20e322aae84c/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.358270    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-k2hwm_6507116d-ad87-4be1-8322-340170f575b3/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.379610    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54ccf4f85d-gbhz2_40210948-0e80-4f69-9d75-ab44c0d38285/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.391159    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-69d6c9f5b8-npvrh_2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.401084    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-jkg9n_1b0c67df-ba69-46b3-ad14-e7287f34fd27/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.419111    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-zwfpg_f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.430236    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-krm9d_59abaeff-0c1b-4f9d-bd55-448a1295f603/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.440153    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5d8f59fb49-q25f6_651ec3cd-3884-41c6-ae9d-1fb08b764a32/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.457231    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-hrzwc_aa122c78-b0a4-4bd5-9844-7a44b150dac4/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.470681    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-wm2vt_cb110aec-f41e-4bae-b0de-9c076a41a603/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.495781    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f_d5b9d0a8-b7a8-4138-a306-4599159f33e1/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.633029    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-698d6bb84b-mv4kv_ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4/operator/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.651617    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-788c8b99b5-fmxvw_12e59f22-9370-4b23-8d8a-5c24ba925541/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.662115    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dlbcw_b85b4390-3574-4e36-bc3a-c0b739646d91/registry-server/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.672815    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-l8kpr_0cf0de8e-51d5-4f14-a63b-08acab6d68a7/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.687000    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-vgkz5_2a362e5c-b09f-4c6c-8639-6c7107e049b0/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.702281    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mhlsp_18d7f488-bb97-4a72-8fbc-eb5db501db16/operator/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.716747    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-ch89p_3370a46a-f9bf-4ce2-977b-6e910b3a4448/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.728986    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-rcxtb_2b9855d9-90b7-4904-9a65-a06a4674d86e/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.739778    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-7wxgh_14b5c974-8804-44d3-9b48-f28f0acab0b6/manager/0.log"
Jan 22 10:03:45 crc kubenswrapper[4824]: I0122 10:03:45.753937    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5ffb9c6597-84f2h_34941af2-d22e-4aaf-b1b6-a620cfc975be/manager/0.log"
Jan 22 10:03:51 crc kubenswrapper[4824]: I0122 10:03:51.703724    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zf5hj_b5902322-e067-4ed7-967d-f71553df3532/control-plane-machine-set-operator/0.log"
Jan 22 10:03:51 crc kubenswrapper[4824]: I0122 10:03:51.720692    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x7psz_8d492ef8-5e48-4fc7-a56b-1233da10e9f7/kube-rbac-proxy/0.log"
Jan 22 10:03:51 crc kubenswrapper[4824]: I0122 10:03:51.731025    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x7psz_8d492ef8-5e48-4fc7-a56b-1233da10e9f7/machine-api-operator/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.287314    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-6khr2_5426de92-d8af-4d6a-9264-64b224018f92/controller/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.293205    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-6khr2_5426de92-d8af-4d6a-9264-64b224018f92/kube-rbac-proxy/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.303856    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-xrhcl_c73c646e-5500-44c3-9799-80a7f7aa6a3c/frr-k8s-webhook-server/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.324294    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/controller/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.506083    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/frr/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.514116    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/reloader/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.518506    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/frr-metrics/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.524795    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/kube-rbac-proxy/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.530992    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/kube-rbac-proxy-frr/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.537032    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-frr-files/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.543191    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-reloader/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.549898    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-metrics/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.571290    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5ff4cd4548-mc6b5_f6f5fc24-b46e-491a-a54d-380666eb1eea/manager/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.580571    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-557b954cc-ldhxs_d51ba000-c226-47ca-a8a8-19d3769e23b3/webhook-server/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.742877    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k77cr_a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad/speaker/0.log"
Jan 22 10:04:01 crc kubenswrapper[4824]: I0122 10:04:01.747414    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k77cr_a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad/kube-rbac-proxy/0.log"
Jan 22 10:04:04 crc kubenswrapper[4824]: I0122 10:04:04.000453    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-9pfww_1d78110c-c6e5-478f-a601-fe69e15d7a1d/cert-manager-controller/0.log"
Jan 22 10:04:04 crc kubenswrapper[4824]: I0122 10:04:04.011187    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-sfrxc_07b9c7cd-b839-415b-ba6a-f3c21c827c7c/cert-manager-cainjector/0.log"
Jan 22 10:04:04 crc kubenswrapper[4824]: I0122 10:04:04.021266    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-mcc4x_4573e1ac-2b18-48bc-93e0-03cf1f11c1b5/cert-manager-webhook/0.log"
Jan 22 10:04:08 crc kubenswrapper[4824]: I0122 10:04:08.906189    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-w7zbs_7a515d93-7907-4474-b22c-cce2da9f6c79/nmstate-console-plugin/0.log"
Jan 22 10:04:08 crc kubenswrapper[4824]: I0122 10:04:08.926993    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2snfz_785191ef-4137-4f5b-808e-640b23400073/nmstate-handler/0.log"
Jan 22 10:04:08 crc kubenswrapper[4824]: I0122 10:04:08.944908    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-x2bnr_9c6e17a7-98cc-4f0f-a999-152a81397e40/nmstate-metrics/0.log"
Jan 22 10:04:08 crc kubenswrapper[4824]: I0122 10:04:08.952761    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-x2bnr_9c6e17a7-98cc-4f0f-a999-152a81397e40/kube-rbac-proxy/0.log"
Jan 22 10:04:08 crc kubenswrapper[4824]: I0122 10:04:08.968336    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-lg7pf_9e097354-ff63-42c7-ae81-c03e10980a87/nmstate-operator/0.log"
Jan 22 10:04:08 crc kubenswrapper[4824]: I0122 10:04:08.981566    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-jqhws_bec9eb1a-8712-4bd6-9e79-684b68436ef8/nmstate-webhook/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.263345    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-6khr2_5426de92-d8af-4d6a-9264-64b224018f92/controller/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.273706    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-6khr2_5426de92-d8af-4d6a-9264-64b224018f92/kube-rbac-proxy/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.286344    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-xrhcl_c73c646e-5500-44c3-9799-80a7f7aa6a3c/frr-k8s-webhook-server/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.306062    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/controller/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.443884    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/frr/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.452785    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/reloader/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.458813    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/frr-metrics/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.466765    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/kube-rbac-proxy/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.480529    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/kube-rbac-proxy-frr/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.484413    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-frr-files/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.492525    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-reloader/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.498387    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-metrics/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.521639    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5ff4cd4548-mc6b5_f6f5fc24-b46e-491a-a54d-380666eb1eea/manager/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.532816    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-557b954cc-ldhxs_d51ba000-c226-47ca-a8a8-19d3769e23b3/webhook-server/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.663831    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k77cr_a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad/speaker/0.log"
Jan 22 10:04:19 crc kubenswrapper[4824]: I0122 10:04:19.672901    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k77cr_a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad/kube-rbac-proxy/0.log"
Jan 22 10:04:23 crc kubenswrapper[4824]: I0122 10:04:23.989375    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf_dfdfcbf5-dbab-420e-a5ee-17c606dba23a/extract/0.log"
Jan 22 10:04:23 crc kubenswrapper[4824]: I0122 10:04:23.997806    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf_dfdfcbf5-dbab-420e-a5ee-17c606dba23a/util/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.029834    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7m4qf_dfdfcbf5-dbab-420e-a5ee-17c606dba23a/pull/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.039094    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w_75984dfb-1883-4993-88f7-7bb43964d1e0/extract/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.046368    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w_75984dfb-1883-4993-88f7-7bb43964d1e0/util/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.055462    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcjn42w_75984dfb-1883-4993-88f7-7bb43964d1e0/pull/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.067430    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4_95a0d0b9-aec9-4fba-89fd-ce6749ada765/extract/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.074263    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4_95a0d0b9-aec9-4fba-89fd-ce6749ada765/util/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.082330    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134xft4_95a0d0b9-aec9-4fba-89fd-ce6749ada765/pull/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.274815    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xgclp_6a59b526-9dc4-423e-a064-d7fd8bef974a/registry-server/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.280943    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xgclp_6a59b526-9dc4-423e-a064-d7fd8bef974a/extract-utilities/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.289977    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-xgclp_6a59b526-9dc4-423e-a064-d7fd8bef974a/extract-content/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.800959    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9ns86_1ed2abbd-725c-4738-9e66-d7143918cb30/registry-server/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.805917    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9ns86_1ed2abbd-725c-4738-9e66-d7143918cb30/extract-utilities/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.813782    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9ns86_1ed2abbd-725c-4738-9e66-d7143918cb30/extract-content/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.831937    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wkknc_08665d12-ff37-4d48-ae5f-caa77a4661d3/marketplace-operator/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.916508    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4wsw_33ac9b68-23da-4599-b9bc-84546dbd1c52/registry-server/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.922380    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4wsw_33ac9b68-23da-4599-b9bc-84546dbd1c52/extract-utilities/0.log"
Jan 22 10:04:24 crc kubenswrapper[4824]: I0122 10:04:24.928215    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c4wsw_33ac9b68-23da-4599-b9bc-84546dbd1c52/extract-content/0.log"
Jan 22 10:04:25 crc kubenswrapper[4824]: I0122 10:04:25.246133    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ksxrb_03cdd9c2-a33d-4b26-adee-57d691ac5eb9/registry-server/0.log"
Jan 22 10:04:25 crc kubenswrapper[4824]: I0122 10:04:25.251202    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ksxrb_03cdd9c2-a33d-4b26-adee-57d691ac5eb9/extract-utilities/0.log"
Jan 22 10:04:25 crc kubenswrapper[4824]: I0122 10:04:25.260671    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ksxrb_03cdd9c2-a33d-4b26-adee-57d691ac5eb9/extract-content/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.856366    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-9pfww_1d78110c-c6e5-478f-a601-fe69e15d7a1d/cert-manager-controller/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.858766    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-6khr2_5426de92-d8af-4d6a-9264-64b224018f92/controller/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.863789    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-6khr2_5426de92-d8af-4d6a-9264-64b224018f92/kube-rbac-proxy/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.875751    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-xrhcl_c73c646e-5500-44c3-9799-80a7f7aa6a3c/frr-k8s-webhook-server/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.884492    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-sfrxc_07b9c7cd-b839-415b-ba6a-f3c21c827c7c/cert-manager-cainjector/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.895163    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-mcc4x_4573e1ac-2b18-48bc-93e0-03cf1f11c1b5/cert-manager-webhook/0.log"
Jan 22 10:05:00 crc kubenswrapper[4824]: I0122 10:05:00.896151    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/controller/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.060319    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/frr/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.067055    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/reloader/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.072559    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/frr-metrics/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.080567    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/kube-rbac-proxy/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.086987    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/kube-rbac-proxy-frr/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.092967    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-frr-files/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.098691    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-reloader/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.108371    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zd6d8_38b26cdc-802e-4435-b3fd-e472393b9c1c/cp-metrics/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.136651    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5ff4cd4548-mc6b5_f6f5fc24-b46e-491a-a54d-380666eb1eea/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.145606    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-557b954cc-ldhxs_d51ba000-c226-47ca-a8a8-19d3769e23b3/webhook-server/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.287371    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k77cr_a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad/speaker/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.294487    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-k77cr_a571f5c4-15c6-42ba-b7e9-353c3ca2e4ad/kube-rbac-proxy/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.766986    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59dd8b7cbf-nbrrg_88e0e1fa-e0ae-43e1-b653-158ea72244b1/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.780134    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-vshwt_d29a7a0c-344a-449a-bfa0-457e860f9030/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.793128    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-8prhs_905ddc8e-1123-4654-92a3-a50493b92a78/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.802532    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/extract/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.812212    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/util/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.823631    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/pull/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.836982    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-rt6zn_b53e7faf-8d7c-4f40-82c7-e6aa64508488/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.850799    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-f5qrl_792e3f5e-1236-4fe0-94b5-20e322aae84c/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.859934    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-k2hwm_6507116d-ad87-4be1-8322-340170f575b3/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.878765    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54ccf4f85d-gbhz2_40210948-0e80-4f69-9d75-ab44c0d38285/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.894319    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-69d6c9f5b8-npvrh_2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.905388    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-jkg9n_1b0c67df-ba69-46b3-ad14-e7287f34fd27/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.918629    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-zwfpg_f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.930172    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-krm9d_59abaeff-0c1b-4f9d-bd55-448a1295f603/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.940962    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5d8f59fb49-q25f6_651ec3cd-3884-41c6-ae9d-1fb08b764a32/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.952314    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-hrzwc_aa122c78-b0a4-4bd5-9844-7a44b150dac4/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.966451    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-wm2vt_cb110aec-f41e-4bae-b0de-9c076a41a603/manager/0.log"
Jan 22 10:05:01 crc kubenswrapper[4824]: I0122 10:05:01.999392    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f_d5b9d0a8-b7a8-4138-a306-4599159f33e1/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.130037    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-698d6bb84b-mv4kv_ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4/operator/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.149012    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-788c8b99b5-fmxvw_12e59f22-9370-4b23-8d8a-5c24ba925541/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.162128    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dlbcw_b85b4390-3574-4e36-bc3a-c0b739646d91/registry-server/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.170133    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-l8kpr_0cf0de8e-51d5-4f14-a63b-08acab6d68a7/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.185470    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-vgkz5_2a362e5c-b09f-4c6c-8639-6c7107e049b0/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.200483    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mhlsp_18d7f488-bb97-4a72-8fbc-eb5db501db16/operator/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.212039    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-ch89p_3370a46a-f9bf-4ce2-977b-6e910b3a4448/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.231034    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-rcxtb_2b9855d9-90b7-4904-9a65-a06a4674d86e/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.244318    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-7wxgh_14b5c974-8804-44d3-9b48-f28f0acab0b6/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.253053    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-9pfww_1d78110c-c6e5-478f-a601-fe69e15d7a1d/cert-manager-controller/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.264662    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5ffb9c6597-84f2h_34941af2-d22e-4aaf-b1b6-a620cfc975be/manager/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.270875    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-sfrxc_07b9c7cd-b839-415b-ba6a-f3c21c827c7c/cert-manager-cainjector/0.log"
Jan 22 10:05:02 crc kubenswrapper[4824]: I0122 10:05:02.384585    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-mcc4x_4573e1ac-2b18-48bc-93e0-03cf1f11c1b5/cert-manager-webhook/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.021696    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zf5hj_b5902322-e067-4ed7-967d-f71553df3532/control-plane-machine-set-operator/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.044786    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x7psz_8d492ef8-5e48-4fc7-a56b-1233da10e9f7/kube-rbac-proxy/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.053086    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x7psz_8d492ef8-5e48-4fc7-a56b-1233da10e9f7/machine-api-operator/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.744943    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-w7zbs_7a515d93-7907-4474-b22c-cce2da9f6c79/nmstate-console-plugin/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.757891    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2snfz_785191ef-4137-4f5b-808e-640b23400073/nmstate-handler/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.771293    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-x2bnr_9c6e17a7-98cc-4f0f-a999-152a81397e40/nmstate-metrics/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.777822    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-x2bnr_9c6e17a7-98cc-4f0f-a999-152a81397e40/kube-rbac-proxy/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.790036    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-lg7pf_9e097354-ff63-42c7-ae81-c03e10980a87/nmstate-operator/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.799874    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-jqhws_bec9eb1a-8712-4bd6-9e79-684b68436ef8/nmstate-webhook/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.832239    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59dd8b7cbf-nbrrg_88e0e1fa-e0ae-43e1-b653-158ea72244b1/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.853159    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-vshwt_d29a7a0c-344a-449a-bfa0-457e860f9030/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.864153    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-8prhs_905ddc8e-1123-4654-92a3-a50493b92a78/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.875164    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/extract/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.882049    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/util/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.889710    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7ace019a3f96c6dacf22bf83e494fa656797b1a183803f1a18d1a3f99mwlt_176eeb37-df5a-41b4-90b6-16b04c0c576c/pull/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.901297    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-rt6zn_b53e7faf-8d7c-4f40-82c7-e6aa64508488/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.911472    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-f5qrl_792e3f5e-1236-4fe0-94b5-20e322aae84c/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.920151    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-k2hwm_6507116d-ad87-4be1-8322-340170f575b3/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.933698    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54ccf4f85d-gbhz2_40210948-0e80-4f69-9d75-ab44c0d38285/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.944228    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-69d6c9f5b8-npvrh_2bb8ede3-1bdc-44ff-a337-3a8b7d2afdf9/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.953514    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-jkg9n_1b0c67df-ba69-46b3-ad14-e7287f34fd27/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.962821    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-zwfpg_f4d0a1a7-c43d-461e-998a-c3bfb7ae0a93/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.973868    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-krm9d_59abaeff-0c1b-4f9d-bd55-448a1295f603/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.984870    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5d8f59fb49-q25f6_651ec3cd-3884-41c6-ae9d-1fb08b764a32/manager/0.log"
Jan 22 10:05:03 crc kubenswrapper[4824]: I0122 10:05:03.996086    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-hrzwc_aa122c78-b0a4-4bd5-9844-7a44b150dac4/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.004807    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-wm2vt_cb110aec-f41e-4bae-b0de-9c076a41a603/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.025738    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c9c58b5572bf7f_d5b9d0a8-b7a8-4138-a306-4599159f33e1/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.159993    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-698d6bb84b-mv4kv_ac2ddd9c-0370-4a81-ae22-8d0e5816ffb4/operator/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.183616    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-788c8b99b5-fmxvw_12e59f22-9370-4b23-8d8a-5c24ba925541/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.196330    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-dlbcw_b85b4390-3574-4e36-bc3a-c0b739646d91/registry-server/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.206447    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-l8kpr_0cf0de8e-51d5-4f14-a63b-08acab6d68a7/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.214921    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-vgkz5_2a362e5c-b09f-4c6c-8639-6c7107e049b0/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.225763    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mhlsp_18d7f488-bb97-4a72-8fbc-eb5db501db16/operator/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.244376    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-ch89p_3370a46a-f9bf-4ce2-977b-6e910b3a4448/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.266417    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-rcxtb_2b9855d9-90b7-4904-9a65-a06a4674d86e/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.276572    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-7wxgh_14b5c974-8804-44d3-9b48-f28f0acab0b6/manager/0.log"
Jan 22 10:05:04 crc kubenswrapper[4824]: I0122 10:05:04.289019    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5ffb9c6597-84f2h_34941af2-d22e-4aaf-b1b6-a620cfc975be/manager/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.913487    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/kube-multus-additional-cni-plugins/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.919210    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/egress-router-binary-copy/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.927055    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/cni-plugins/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.933035    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/bond-cni-plugin/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.938700    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/routeoverride-cni/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.944959    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/whereabouts-cni-bincopy/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.951163    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-9wczc_68893d7c-efac-44f9-8935-843169e413ad/whereabouts-cni/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.965104    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-hpp2x_0e085a8e-5760-4859-bf6b-ac4fac6aa9cf/multus-admission-controller/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.969099    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-hpp2x_0e085a8e-5760-4859-bf6b-ac4fac6aa9cf/kube-rbac-proxy/0.log"
Jan 22 10:05:05 crc kubenswrapper[4824]: I0122 10:05:05.993691    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/3.log"
Jan 22 10:05:06 crc kubenswrapper[4824]: I0122 10:05:06.010251    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-zvvt5_00f6a0cb-f47d-41ea-8cac-ca5454ce7c0c/kube-multus/2.log"
Jan 22 10:05:06 crc kubenswrapper[4824]: I0122 10:05:06.029859    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-xxtjz_f43b4618-291a-40d3-b1ca-32abfadaf376/network-metrics-daemon/0.log"
Jan 22 10:05:06 crc kubenswrapper[4824]: I0122 10:05:06.034107    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-xxtjz_f43b4618-291a-40d3-b1ca-32abfadaf376/kube-rbac-proxy/0.log"
Jan 22 10:06:00 crc kubenswrapper[4824]: I0122 10:06:00.333111    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:06:00 crc kubenswrapper[4824]: I0122 10:06:00.335293    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:06:30 crc kubenswrapper[4824]: I0122 10:06:30.333426    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:06:30 crc kubenswrapper[4824]: I0122 10:06:30.333907    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:07:00 crc kubenswrapper[4824]: I0122 10:07:00.332919    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:07:00 crc kubenswrapper[4824]: I0122 10:07:00.333469    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:07:00 crc kubenswrapper[4824]: I0122 10:07:00.333508    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 10:07:00 crc kubenswrapper[4824]: I0122 10:07:00.334073    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"363a321edd4efcce621b593f7e6704f72529b44f15f265b34c10f2ff6d4ad874"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 10:07:00 crc kubenswrapper[4824]: I0122 10:07:00.334174    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://363a321edd4efcce621b593f7e6704f72529b44f15f265b34c10f2ff6d4ad874" gracePeriod=600
Jan 22 10:07:01 crc kubenswrapper[4824]: I0122 10:07:01.553967    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="363a321edd4efcce621b593f7e6704f72529b44f15f265b34c10f2ff6d4ad874" exitCode=0
Jan 22 10:07:01 crc kubenswrapper[4824]: I0122 10:07:01.554057    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"363a321edd4efcce621b593f7e6704f72529b44f15f265b34c10f2ff6d4ad874"}
Jan 22 10:07:01 crc kubenswrapper[4824]: I0122 10:07:01.554566    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"}
Jan 22 10:07:01 crc kubenswrapper[4824]: I0122 10:07:01.554585    4824 scope.go:117] "RemoveContainer" containerID="142d82198208020ed04e19461fd1e7bda1b5d653c554ec42bffda74db5a4d02d"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.149409    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5zg5k"]
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.152600    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.164468    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zg5k"]
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.323329    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-catalog-content\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.323865    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-utilities\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.324058    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxlz9\" (UniqueName: \"kubernetes.io/projected/373d2368-c139-4537-a6f5-9b051b8395d0-kube-api-access-wxlz9\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.425707    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxlz9\" (UniqueName: \"kubernetes.io/projected/373d2368-c139-4537-a6f5-9b051b8395d0-kube-api-access-wxlz9\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.425801    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-catalog-content\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.425835    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-utilities\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.426415    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-utilities\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.426489    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-catalog-content\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.451805    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxlz9\" (UniqueName: \"kubernetes.io/projected/373d2368-c139-4537-a6f5-9b051b8395d0-kube-api-access-wxlz9\") pod \"community-operators-5zg5k\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") " pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.473421    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.773161    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5zg5k"]
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.942139    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerStarted","Data":"453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf"}
Jan 22 10:07:50 crc kubenswrapper[4824]: I0122 10:07:50.942178    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerStarted","Data":"52ebad1a70c88282d2ed193b48ee994c296815b70e98def3acec2df9a20f5d81"}
Jan 22 10:07:51 crc kubenswrapper[4824]: I0122 10:07:51.951895    4824 generic.go:334] "Generic (PLEG): container finished" podID="373d2368-c139-4537-a6f5-9b051b8395d0" containerID="453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf" exitCode=0
Jan 22 10:07:51 crc kubenswrapper[4824]: I0122 10:07:51.952068    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerDied","Data":"453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf"}
Jan 22 10:07:51 crc kubenswrapper[4824]: I0122 10:07:51.954798    4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Jan 22 10:07:52 crc kubenswrapper[4824]: I0122 10:07:52.962404    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerStarted","Data":"7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f"}
Jan 22 10:07:53 crc kubenswrapper[4824]: I0122 10:07:53.977722    4824 generic.go:334] "Generic (PLEG): container finished" podID="373d2368-c139-4537-a6f5-9b051b8395d0" containerID="7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f" exitCode=0
Jan 22 10:07:53 crc kubenswrapper[4824]: I0122 10:07:53.978141    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerDied","Data":"7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f"}
Jan 22 10:07:54 crc kubenswrapper[4824]: I0122 10:07:54.988249    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerStarted","Data":"70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d"}
Jan 22 10:07:55 crc kubenswrapper[4824]: I0122 10:07:55.013698    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5zg5k" podStartSLOduration=2.59708503 podStartE2EDuration="5.013680612s" podCreationTimestamp="2026-01-22 10:07:50 +0000 UTC" firstStartedPulling="2026-01-22 10:07:51.954489511 +0000 UTC m=+1504.857966514" lastFinishedPulling="2026-01-22 10:07:54.371085103 +0000 UTC m=+1507.274562096" observedRunningTime="2026-01-22 10:07:55.009312812 +0000 UTC m=+1507.912789815" watchObservedRunningTime="2026-01-22 10:07:55.013680612 +0000 UTC m=+1507.917157605"
Jan 22 10:08:00 crc kubenswrapper[4824]: I0122 10:08:00.473993    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:08:00 crc kubenswrapper[4824]: I0122 10:08:00.474782    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:08:00 crc kubenswrapper[4824]: I0122 10:08:00.550372    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:08:01 crc kubenswrapper[4824]: I0122 10:08:01.116350    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:08:01 crc kubenswrapper[4824]: I0122 10:08:01.173792    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zg5k"]
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.083046    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5zg5k" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="registry-server" containerID="cri-o://70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d" gracePeriod=2
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.714566    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.909809    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-utilities\") pod \"373d2368-c139-4537-a6f5-9b051b8395d0\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") "
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.909918    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-catalog-content\") pod \"373d2368-c139-4537-a6f5-9b051b8395d0\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") "
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.909968    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxlz9\" (UniqueName: \"kubernetes.io/projected/373d2368-c139-4537-a6f5-9b051b8395d0-kube-api-access-wxlz9\") pod \"373d2368-c139-4537-a6f5-9b051b8395d0\" (UID: \"373d2368-c139-4537-a6f5-9b051b8395d0\") "
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.910729    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-utilities" (OuterVolumeSpecName: "utilities") pod "373d2368-c139-4537-a6f5-9b051b8395d0" (UID: "373d2368-c139-4537-a6f5-9b051b8395d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.917315    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373d2368-c139-4537-a6f5-9b051b8395d0-kube-api-access-wxlz9" (OuterVolumeSpecName: "kube-api-access-wxlz9") pod "373d2368-c139-4537-a6f5-9b051b8395d0" (UID: "373d2368-c139-4537-a6f5-9b051b8395d0"). InnerVolumeSpecName "kube-api-access-wxlz9". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:08:03 crc kubenswrapper[4824]: I0122 10:08:03.990679    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "373d2368-c139-4537-a6f5-9b051b8395d0" (UID: "373d2368-c139-4537-a6f5-9b051b8395d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.012102    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.012172    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxlz9\" (UniqueName: \"kubernetes.io/projected/373d2368-c139-4537-a6f5-9b051b8395d0-kube-api-access-wxlz9\") on node \"crc\" DevicePath \"\""
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.012188    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/373d2368-c139-4537-a6f5-9b051b8395d0-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.092108    4824 generic.go:334] "Generic (PLEG): container finished" podID="373d2368-c139-4537-a6f5-9b051b8395d0" containerID="70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d" exitCode=0
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.092167    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerDied","Data":"70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d"}
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.092191    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5zg5k"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.092222    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5zg5k" event={"ID":"373d2368-c139-4537-a6f5-9b051b8395d0","Type":"ContainerDied","Data":"52ebad1a70c88282d2ed193b48ee994c296815b70e98def3acec2df9a20f5d81"}
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.092240    4824 scope.go:117] "RemoveContainer" containerID="70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.115744    4824 scope.go:117] "RemoveContainer" containerID="7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.134101    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5zg5k"]
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.139209    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5zg5k"]
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.146199    4824 scope.go:117] "RemoveContainer" containerID="453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.166238    4824 scope.go:117] "RemoveContainer" containerID="70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d"
Jan 22 10:08:04 crc kubenswrapper[4824]: E0122 10:08:04.166680    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d\": container with ID starting with 70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d not found: ID does not exist" containerID="70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.166708    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d"} err="failed to get container status \"70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d\": rpc error: code = NotFound desc = could not find container \"70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d\": container with ID starting with 70bcadcfa0a1eafa23ec68b5ea26b8d35974fd2014242f4fcfef4a23bbdb832d not found: ID does not exist"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.166730    4824 scope.go:117] "RemoveContainer" containerID="7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f"
Jan 22 10:08:04 crc kubenswrapper[4824]: E0122 10:08:04.167220    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f\": container with ID starting with 7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f not found: ID does not exist" containerID="7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.167248    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f"} err="failed to get container status \"7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f\": rpc error: code = NotFound desc = could not find container \"7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f\": container with ID starting with 7d610a4ffbca95e93f8112f7af9bd647418b53490d26796d36d8e0631dc64f9f not found: ID does not exist"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.167266    4824 scope.go:117] "RemoveContainer" containerID="453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf"
Jan 22 10:08:04 crc kubenswrapper[4824]: E0122 10:08:04.167716    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf\": container with ID starting with 453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf not found: ID does not exist" containerID="453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf"
Jan 22 10:08:04 crc kubenswrapper[4824]: I0122 10:08:04.167755    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf"} err="failed to get container status \"453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf\": rpc error: code = NotFound desc = could not find container \"453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf\": container with ID starting with 453e444554b5f28a18a0539631d7e87585d59153f2ed7bd3d52194010aaae1cf not found: ID does not exist"
Jan 22 10:08:05 crc kubenswrapper[4824]: I0122 10:08:05.417079    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" path="/var/lib/kubelet/pods/373d2368-c139-4537-a6f5-9b051b8395d0/volumes"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.299012    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x4clw"]
Jan 22 10:08:15 crc kubenswrapper[4824]: E0122 10:08:15.299845    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="extract-content"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.299884    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="extract-content"
Jan 22 10:08:15 crc kubenswrapper[4824]: E0122 10:08:15.299911    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="registry-server"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.299923    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="registry-server"
Jan 22 10:08:15 crc kubenswrapper[4824]: E0122 10:08:15.299955    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="extract-utilities"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.299970    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="extract-utilities"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.300231    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="373d2368-c139-4537-a6f5-9b051b8395d0" containerName="registry-server"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.302194    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.303513    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-utilities\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.303579    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-catalog-content\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.303603    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq4zr\" (UniqueName: \"kubernetes.io/projected/0523073f-2480-4469-aa30-6eb7a2c4d384-kube-api-access-nq4zr\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.304710    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4clw"]
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.404651    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq4zr\" (UniqueName: \"kubernetes.io/projected/0523073f-2480-4469-aa30-6eb7a2c4d384-kube-api-access-nq4zr\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.405086    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-utilities\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.405193    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-catalog-content\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.405625    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-catalog-content\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.405943    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-utilities\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.440405    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq4zr\" (UniqueName: \"kubernetes.io/projected/0523073f-2480-4469-aa30-6eb7a2c4d384-kube-api-access-nq4zr\") pod \"redhat-marketplace-x4clw\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") " pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:15 crc kubenswrapper[4824]: I0122 10:08:15.620752    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:16 crc kubenswrapper[4824]: I0122 10:08:16.124571    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4clw"]
Jan 22 10:08:16 crc kubenswrapper[4824]: I0122 10:08:16.193723    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4clw" event={"ID":"0523073f-2480-4469-aa30-6eb7a2c4d384","Type":"ContainerStarted","Data":"01670cd61ca4c86a783f61b857ca51c77b42c8fdad1a3438efa4cc74b0afd1da"}
Jan 22 10:08:17 crc kubenswrapper[4824]: I0122 10:08:17.211516    4824 generic.go:334] "Generic (PLEG): container finished" podID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerID="00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd" exitCode=0
Jan 22 10:08:17 crc kubenswrapper[4824]: I0122 10:08:17.211557    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4clw" event={"ID":"0523073f-2480-4469-aa30-6eb7a2c4d384","Type":"ContainerDied","Data":"00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd"}
Jan 22 10:08:18 crc kubenswrapper[4824]: I0122 10:08:18.221086    4824 generic.go:334] "Generic (PLEG): container finished" podID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerID="06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace" exitCode=0
Jan 22 10:08:18 crc kubenswrapper[4824]: I0122 10:08:18.221156    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4clw" event={"ID":"0523073f-2480-4469-aa30-6eb7a2c4d384","Type":"ContainerDied","Data":"06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace"}
Jan 22 10:08:19 crc kubenswrapper[4824]: I0122 10:08:19.232880    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4clw" event={"ID":"0523073f-2480-4469-aa30-6eb7a2c4d384","Type":"ContainerStarted","Data":"9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549"}
Jan 22 10:08:19 crc kubenswrapper[4824]: I0122 10:08:19.257454    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x4clw" podStartSLOduration=2.7032488260000003 podStartE2EDuration="4.257427949s" podCreationTimestamp="2026-01-22 10:08:15 +0000 UTC" firstStartedPulling="2026-01-22 10:08:17.213392397 +0000 UTC m=+1530.116869390" lastFinishedPulling="2026-01-22 10:08:18.76757151 +0000 UTC m=+1531.671048513" observedRunningTime="2026-01-22 10:08:19.251436308 +0000 UTC m=+1532.154913331" watchObservedRunningTime="2026-01-22 10:08:19.257427949 +0000 UTC m=+1532.160904982"
Jan 22 10:08:25 crc kubenswrapper[4824]: I0122 10:08:25.622027    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:25 crc kubenswrapper[4824]: I0122 10:08:25.623547    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:25 crc kubenswrapper[4824]: I0122 10:08:25.670123    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:26 crc kubenswrapper[4824]: I0122 10:08:26.316567    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:26 crc kubenswrapper[4824]: I0122 10:08:26.362112    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4clw"]
Jan 22 10:08:28 crc kubenswrapper[4824]: I0122 10:08:28.287896    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x4clw" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="registry-server" containerID="cri-o://9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549" gracePeriod=2
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.246948    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.294291    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq4zr\" (UniqueName: \"kubernetes.io/projected/0523073f-2480-4469-aa30-6eb7a2c4d384-kube-api-access-nq4zr\") pod \"0523073f-2480-4469-aa30-6eb7a2c4d384\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") "
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.294384    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-catalog-content\") pod \"0523073f-2480-4469-aa30-6eb7a2c4d384\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") "
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.294418    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-utilities\") pod \"0523073f-2480-4469-aa30-6eb7a2c4d384\" (UID: \"0523073f-2480-4469-aa30-6eb7a2c4d384\") "
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.296696    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-utilities" (OuterVolumeSpecName: "utilities") pod "0523073f-2480-4469-aa30-6eb7a2c4d384" (UID: "0523073f-2480-4469-aa30-6eb7a2c4d384"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.299645    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x4clw"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.299667    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4clw" event={"ID":"0523073f-2480-4469-aa30-6eb7a2c4d384","Type":"ContainerDied","Data":"9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549"}
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.299718    4824 scope.go:117] "RemoveContainer" containerID="9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.299586    4824 generic.go:334] "Generic (PLEG): container finished" podID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerID="9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549" exitCode=0
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.299778    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x4clw" event={"ID":"0523073f-2480-4469-aa30-6eb7a2c4d384","Type":"ContainerDied","Data":"01670cd61ca4c86a783f61b857ca51c77b42c8fdad1a3438efa4cc74b0afd1da"}
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.316643    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0523073f-2480-4469-aa30-6eb7a2c4d384" (UID: "0523073f-2480-4469-aa30-6eb7a2c4d384"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.318160    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0523073f-2480-4469-aa30-6eb7a2c4d384-kube-api-access-nq4zr" (OuterVolumeSpecName: "kube-api-access-nq4zr") pod "0523073f-2480-4469-aa30-6eb7a2c4d384" (UID: "0523073f-2480-4469-aa30-6eb7a2c4d384"). InnerVolumeSpecName "kube-api-access-nq4zr". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.319976    4824 scope.go:117] "RemoveContainer" containerID="06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.350413    4824 scope.go:117] "RemoveContainer" containerID="00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.378320    4824 scope.go:117] "RemoveContainer" containerID="9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549"
Jan 22 10:08:29 crc kubenswrapper[4824]: E0122 10:08:29.378847    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549\": container with ID starting with 9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549 not found: ID does not exist" containerID="9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.378876    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549"} err="failed to get container status \"9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549\": rpc error: code = NotFound desc = could not find container \"9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549\": container with ID starting with 9a4728449412acec35c785b247d277891fa37c5ad7d59663f6efdd654a2b7549 not found: ID does not exist"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.378899    4824 scope.go:117] "RemoveContainer" containerID="06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace"
Jan 22 10:08:29 crc kubenswrapper[4824]: E0122 10:08:29.379184    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace\": container with ID starting with 06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace not found: ID does not exist" containerID="06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.379212    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace"} err="failed to get container status \"06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace\": rpc error: code = NotFound desc = could not find container \"06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace\": container with ID starting with 06645d0964d587f9131af39ea89c657ac96e67fc32f40c04de5fac7b907e1ace not found: ID does not exist"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.379227    4824 scope.go:117] "RemoveContainer" containerID="00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd"
Jan 22 10:08:29 crc kubenswrapper[4824]: E0122 10:08:29.379490    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd\": container with ID starting with 00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd not found: ID does not exist" containerID="00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.379512    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd"} err="failed to get container status \"00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd\": rpc error: code = NotFound desc = could not find container \"00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd\": container with ID starting with 00540bf9ab5a218fee33a22ebea87ccc68d9ff26e5520cd8075407776b8ca6cd not found: ID does not exist"
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.395490    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq4zr\" (UniqueName: \"kubernetes.io/projected/0523073f-2480-4469-aa30-6eb7a2c4d384-kube-api-access-nq4zr\") on node \"crc\" DevicePath \"\""
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.395510    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.395519    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0523073f-2480-4469-aa30-6eb7a2c4d384-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.624649    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4clw"]
Jan 22 10:08:29 crc kubenswrapper[4824]: I0122 10:08:29.631915    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x4clw"]
Jan 22 10:08:31 crc kubenswrapper[4824]: I0122 10:08:31.414693    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" path="/var/lib/kubelet/pods/0523073f-2480-4469-aa30-6eb7a2c4d384/volumes"
Jan 22 10:09:30 crc kubenswrapper[4824]: I0122 10:09:30.333097    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:09:30 crc kubenswrapper[4824]: I0122 10:09:30.335430    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:10:00 crc kubenswrapper[4824]: I0122 10:10:00.332534    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:10:00 crc kubenswrapper[4824]: I0122 10:10:00.333095    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:10:30 crc kubenswrapper[4824]: I0122 10:10:30.332764    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:10:30 crc kubenswrapper[4824]: I0122 10:10:30.333652    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:10:30 crc kubenswrapper[4824]: I0122 10:10:30.333712    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 10:10:30 crc kubenswrapper[4824]: I0122 10:10:30.334589    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 10:10:30 crc kubenswrapper[4824]: I0122 10:10:30.334641    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15" gracePeriod=600
Jan 22 10:10:30 crc kubenswrapper[4824]: E0122 10:10:30.456285    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:10:31 crc kubenswrapper[4824]: I0122 10:10:31.200385    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15" exitCode=0
Jan 22 10:10:31 crc kubenswrapper[4824]: I0122 10:10:31.200439    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"}
Jan 22 10:10:31 crc kubenswrapper[4824]: I0122 10:10:31.200478    4824 scope.go:117] "RemoveContainer" containerID="363a321edd4efcce621b593f7e6704f72529b44f15f265b34c10f2ff6d4ad874"
Jan 22 10:10:31 crc kubenswrapper[4824]: I0122 10:10:31.201021    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:10:31 crc kubenswrapper[4824]: E0122 10:10:31.201246    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:10:44 crc kubenswrapper[4824]: I0122 10:10:44.404247    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:10:44 crc kubenswrapper[4824]: E0122 10:10:44.404877    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:10:57 crc kubenswrapper[4824]: I0122 10:10:57.408026    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:10:57 crc kubenswrapper[4824]: E0122 10:10:57.408840    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:11:10 crc kubenswrapper[4824]: I0122 10:11:10.405146    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:11:10 crc kubenswrapper[4824]: E0122 10:11:10.406472    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:11:23 crc kubenswrapper[4824]: I0122 10:11:23.404093    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:11:23 crc kubenswrapper[4824]: E0122 10:11:23.404837    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:11:36 crc kubenswrapper[4824]: I0122 10:11:36.404639    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:11:36 crc kubenswrapper[4824]: E0122 10:11:36.407758    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:11:48 crc kubenswrapper[4824]: I0122 10:11:48.403559    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:11:48 crc kubenswrapper[4824]: E0122 10:11:48.404163    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:12:02 crc kubenswrapper[4824]: I0122 10:12:02.404709    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:12:02 crc kubenswrapper[4824]: E0122 10:12:02.405979    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:12:14 crc kubenswrapper[4824]: I0122 10:12:14.404987    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:12:14 crc kubenswrapper[4824]: E0122 10:12:14.406608    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:12:26 crc kubenswrapper[4824]: I0122 10:12:26.403933    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:12:26 crc kubenswrapper[4824]: E0122 10:12:26.404655    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:12:41 crc kubenswrapper[4824]: I0122 10:12:41.404725    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:12:41 crc kubenswrapper[4824]: E0122 10:12:41.405449    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:12:55 crc kubenswrapper[4824]: I0122 10:12:55.406326    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:12:55 crc kubenswrapper[4824]: E0122 10:12:55.407174    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:13:07 crc kubenswrapper[4824]: I0122 10:13:07.409525    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:13:07 crc kubenswrapper[4824]: E0122 10:13:07.410513    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:13:20 crc kubenswrapper[4824]: I0122 10:13:20.404301    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:13:20 crc kubenswrapper[4824]: E0122 10:13:20.404914    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:13:34 crc kubenswrapper[4824]: I0122 10:13:34.403973    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:13:34 crc kubenswrapper[4824]: E0122 10:13:34.404990    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:13:47 crc kubenswrapper[4824]: I0122 10:13:47.419839    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:13:47 crc kubenswrapper[4824]: E0122 10:13:47.420920    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:14:02 crc kubenswrapper[4824]: I0122 10:14:02.404703    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:14:02 crc kubenswrapper[4824]: E0122 10:14:02.406673    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:14:15 crc kubenswrapper[4824]: I0122 10:14:15.414865    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:14:15 crc kubenswrapper[4824]: E0122 10:14:15.416144    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:14:28 crc kubenswrapper[4824]: I0122 10:14:28.404008    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:14:28 crc kubenswrapper[4824]: E0122 10:14:28.404874    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:14:40 crc kubenswrapper[4824]: I0122 10:14:40.404450    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:14:40 crc kubenswrapper[4824]: E0122 10:14:40.405141    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:14:52 crc kubenswrapper[4824]: I0122 10:14:52.403891    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:14:52 crc kubenswrapper[4824]: E0122 10:14:52.404573    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.161023    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"]
Jan 22 10:15:00 crc kubenswrapper[4824]: E0122 10:15:00.162278    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="extract-utilities"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.162296    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="extract-utilities"
Jan 22 10:15:00 crc kubenswrapper[4824]: E0122 10:15:00.162310    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="extract-content"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.162316    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="extract-content"
Jan 22 10:15:00 crc kubenswrapper[4824]: E0122 10:15:00.162345    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="registry-server"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.162418    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="registry-server"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.162584    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0523073f-2480-4469-aa30-6eb7a2c4d384" containerName="registry-server"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.163241    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.166815    4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.167578    4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.167966    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e5995b3-17d8-4318-8e1b-5effa9322520-secret-volume\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.168144    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr6hx\" (UniqueName: \"kubernetes.io/projected/0e5995b3-17d8-4318-8e1b-5effa9322520-kube-api-access-hr6hx\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.168242    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e5995b3-17d8-4318-8e1b-5effa9322520-config-volume\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.172381    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"]
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.270369    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e5995b3-17d8-4318-8e1b-5effa9322520-config-volume\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.270501    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e5995b3-17d8-4318-8e1b-5effa9322520-secret-volume\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.270560    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr6hx\" (UniqueName: \"kubernetes.io/projected/0e5995b3-17d8-4318-8e1b-5effa9322520-kube-api-access-hr6hx\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.271691    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e5995b3-17d8-4318-8e1b-5effa9322520-config-volume\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.285186    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e5995b3-17d8-4318-8e1b-5effa9322520-secret-volume\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.285746    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr6hx\" (UniqueName: \"kubernetes.io/projected/0e5995b3-17d8-4318-8e1b-5effa9322520-kube-api-access-hr6hx\") pod \"collect-profiles-29484615-7nxvw\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:00 crc kubenswrapper[4824]: I0122 10:15:00.483605    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:01 crc kubenswrapper[4824]: I0122 10:15:01.096481    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"]
Jan 22 10:15:01 crc kubenswrapper[4824]: I0122 10:15:01.482444    4824 generic.go:334] "Generic (PLEG): container finished" podID="0e5995b3-17d8-4318-8e1b-5effa9322520" containerID="0850574dceae5d0b2c3f7d3d6c7b919601f97c3ce2ef6a928277c7e15e75768f" exitCode=0
Jan 22 10:15:01 crc kubenswrapper[4824]: I0122 10:15:01.482487    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw" event={"ID":"0e5995b3-17d8-4318-8e1b-5effa9322520","Type":"ContainerDied","Data":"0850574dceae5d0b2c3f7d3d6c7b919601f97c3ce2ef6a928277c7e15e75768f"}
Jan 22 10:15:01 crc kubenswrapper[4824]: I0122 10:15:01.482510    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw" event={"ID":"0e5995b3-17d8-4318-8e1b-5effa9322520","Type":"ContainerStarted","Data":"6fa3379fb08cfb0ba3cfc439b7e8cf94664b9e4b53efb3c9233a98fe90e2fa28"}
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.799940    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.811203    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e5995b3-17d8-4318-8e1b-5effa9322520-secret-volume\") pod \"0e5995b3-17d8-4318-8e1b-5effa9322520\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") "
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.811283    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e5995b3-17d8-4318-8e1b-5effa9322520-config-volume\") pod \"0e5995b3-17d8-4318-8e1b-5effa9322520\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") "
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.811313    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr6hx\" (UniqueName: \"kubernetes.io/projected/0e5995b3-17d8-4318-8e1b-5effa9322520-kube-api-access-hr6hx\") pod \"0e5995b3-17d8-4318-8e1b-5effa9322520\" (UID: \"0e5995b3-17d8-4318-8e1b-5effa9322520\") "
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.812163    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e5995b3-17d8-4318-8e1b-5effa9322520-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e5995b3-17d8-4318-8e1b-5effa9322520" (UID: "0e5995b3-17d8-4318-8e1b-5effa9322520"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.818280    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e5995b3-17d8-4318-8e1b-5effa9322520-kube-api-access-hr6hx" (OuterVolumeSpecName: "kube-api-access-hr6hx") pod "0e5995b3-17d8-4318-8e1b-5effa9322520" (UID: "0e5995b3-17d8-4318-8e1b-5effa9322520"). InnerVolumeSpecName "kube-api-access-hr6hx". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.822459    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e5995b3-17d8-4318-8e1b-5effa9322520-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e5995b3-17d8-4318-8e1b-5effa9322520" (UID: "0e5995b3-17d8-4318-8e1b-5effa9322520"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.912340    4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e5995b3-17d8-4318-8e1b-5effa9322520-secret-volume\") on node \"crc\" DevicePath \"\""
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.912763    4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e5995b3-17d8-4318-8e1b-5effa9322520-config-volume\") on node \"crc\" DevicePath \"\""
Jan 22 10:15:02 crc kubenswrapper[4824]: I0122 10:15:02.912777    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr6hx\" (UniqueName: \"kubernetes.io/projected/0e5995b3-17d8-4318-8e1b-5effa9322520-kube-api-access-hr6hx\") on node \"crc\" DevicePath \"\""
Jan 22 10:15:03 crc kubenswrapper[4824]: I0122 10:15:03.497032    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw" event={"ID":"0e5995b3-17d8-4318-8e1b-5effa9322520","Type":"ContainerDied","Data":"6fa3379fb08cfb0ba3cfc439b7e8cf94664b9e4b53efb3c9233a98fe90e2fa28"}
Jan 22 10:15:03 crc kubenswrapper[4824]: I0122 10:15:03.497066    4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fa3379fb08cfb0ba3cfc439b7e8cf94664b9e4b53efb3c9233a98fe90e2fa28"
Jan 22 10:15:03 crc kubenswrapper[4824]: I0122 10:15:03.497145    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29484615-7nxvw"
Jan 22 10:15:05 crc kubenswrapper[4824]: I0122 10:15:05.404518    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:15:05 crc kubenswrapper[4824]: E0122 10:15:05.404895    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:15:18 crc kubenswrapper[4824]: I0122 10:15:18.404800    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:15:18 crc kubenswrapper[4824]: E0122 10:15:18.405769    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:15:32 crc kubenswrapper[4824]: I0122 10:15:32.405038    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:15:32 crc kubenswrapper[4824]: I0122 10:15:32.739767    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"2eaeaf473f7555a3f6d348915a2aedbf1a90414aeb0df2817d91de3ec90015a3"}
Jan 22 10:15:58 crc kubenswrapper[4824]: I0122 10:15:58.973672    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nfsgl"]
Jan 22 10:15:58 crc kubenswrapper[4824]: E0122 10:15:58.987849    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e5995b3-17d8-4318-8e1b-5effa9322520" containerName="collect-profiles"
Jan 22 10:15:58 crc kubenswrapper[4824]: I0122 10:15:58.988116    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e5995b3-17d8-4318-8e1b-5effa9322520" containerName="collect-profiles"
Jan 22 10:15:58 crc kubenswrapper[4824]: I0122 10:15:58.990009    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e5995b3-17d8-4318-8e1b-5effa9322520" containerName="collect-profiles"
Jan 22 10:15:58 crc kubenswrapper[4824]: I0122 10:15:58.998509    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nfsgl"]
Jan 22 10:15:58 crc kubenswrapper[4824]: I0122 10:15:58.998710    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.036627    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-catalog-content\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.036745    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-utilities\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.036787    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpk8v\" (UniqueName: \"kubernetes.io/projected/68ff50fc-c751-42f3-8d84-cbd701e52035-kube-api-access-kpk8v\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.138470    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpk8v\" (UniqueName: \"kubernetes.io/projected/68ff50fc-c751-42f3-8d84-cbd701e52035-kube-api-access-kpk8v\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.138560    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-catalog-content\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.138622    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-utilities\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.139037    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-utilities\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.139212    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-catalog-content\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.158685    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpk8v\" (UniqueName: \"kubernetes.io/projected/68ff50fc-c751-42f3-8d84-cbd701e52035-kube-api-access-kpk8v\") pod \"certified-operators-nfsgl\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") " pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.318902    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.831263    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nfsgl"]
Jan 22 10:15:59 crc kubenswrapper[4824]: I0122 10:15:59.949603    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfsgl" event={"ID":"68ff50fc-c751-42f3-8d84-cbd701e52035","Type":"ContainerStarted","Data":"85d87d444edd0b5c8f42d087aeb9cf596c87d5eb1dd476e1635f70c4ea369031"}
Jan 22 10:16:00 crc kubenswrapper[4824]: I0122 10:16:00.962442    4824 generic.go:334] "Generic (PLEG): container finished" podID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerID="9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801" exitCode=0
Jan 22 10:16:00 crc kubenswrapper[4824]: I0122 10:16:00.962509    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfsgl" event={"ID":"68ff50fc-c751-42f3-8d84-cbd701e52035","Type":"ContainerDied","Data":"9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801"}
Jan 22 10:16:00 crc kubenswrapper[4824]: I0122 10:16:00.965590    4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider
Jan 22 10:16:02 crc kubenswrapper[4824]: I0122 10:16:02.980840    4824 generic.go:334] "Generic (PLEG): container finished" podID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerID="ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b" exitCode=0
Jan 22 10:16:02 crc kubenswrapper[4824]: I0122 10:16:02.981089    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfsgl" event={"ID":"68ff50fc-c751-42f3-8d84-cbd701e52035","Type":"ContainerDied","Data":"ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b"}
Jan 22 10:16:03 crc kubenswrapper[4824]: I0122 10:16:03.991928    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfsgl" event={"ID":"68ff50fc-c751-42f3-8d84-cbd701e52035","Type":"ContainerStarted","Data":"9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549"}
Jan 22 10:16:04 crc kubenswrapper[4824]: I0122 10:16:04.022911    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nfsgl" podStartSLOduration=3.281088705 podStartE2EDuration="6.022895668s" podCreationTimestamp="2026-01-22 10:15:58 +0000 UTC" firstStartedPulling="2026-01-22 10:16:00.965176164 +0000 UTC m=+1993.868653167" lastFinishedPulling="2026-01-22 10:16:03.706983137 +0000 UTC m=+1996.610460130" observedRunningTime="2026-01-22 10:16:04.016149478 +0000 UTC m=+1996.919626471" watchObservedRunningTime="2026-01-22 10:16:04.022895668 +0000 UTC m=+1996.926372661"
Jan 22 10:16:09 crc kubenswrapper[4824]: I0122 10:16:09.320512    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:16:09 crc kubenswrapper[4824]: I0122 10:16:09.322411    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:16:09 crc kubenswrapper[4824]: I0122 10:16:09.367342    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:16:10 crc kubenswrapper[4824]: I0122 10:16:10.086922    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:16:12 crc kubenswrapper[4824]: I0122 10:16:12.813570    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dqhv9"]
Jan 22 10:16:12 crc kubenswrapper[4824]: I0122 10:16:12.815655    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:12 crc kubenswrapper[4824]: I0122 10:16:12.830037    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqhv9"]
Jan 22 10:16:12 crc kubenswrapper[4824]: I0122 10:16:12.952200    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-catalog-content\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:12 crc kubenswrapper[4824]: I0122 10:16:12.952273    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4q2q\" (UniqueName: \"kubernetes.io/projected/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-kube-api-access-k4q2q\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:12 crc kubenswrapper[4824]: I0122 10:16:12.952323    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-utilities\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.054103    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-utilities\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.054186    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-catalog-content\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.054220    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4q2q\" (UniqueName: \"kubernetes.io/projected/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-kube-api-access-k4q2q\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.054689    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-utilities\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.054744    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-catalog-content\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.075440    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4q2q\" (UniqueName: \"kubernetes.io/projected/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-kube-api-access-k4q2q\") pod \"redhat-operators-dqhv9\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") " pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.132274    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:13 crc kubenswrapper[4824]: I0122 10:16:13.361884    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dqhv9"]
Jan 22 10:16:14 crc kubenswrapper[4824]: I0122 10:16:14.066939    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerStarted","Data":"81c5ee8a3f4444d210e186e4b1e7027f2bdf6561db5286fd13f75c8729283060"}
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.076340    4824 generic.go:334] "Generic (PLEG): container finished" podID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerID="da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0" exitCode=0
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.076414    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerDied","Data":"da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0"}
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.209219    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nfsgl"]
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.209503    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nfsgl" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="registry-server" containerID="cri-o://9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549" gracePeriod=2
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.648778    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.714641    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-catalog-content\") pod \"68ff50fc-c751-42f3-8d84-cbd701e52035\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") "
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.783496    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68ff50fc-c751-42f3-8d84-cbd701e52035" (UID: "68ff50fc-c751-42f3-8d84-cbd701e52035"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.816017    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-utilities\") pod \"68ff50fc-c751-42f3-8d84-cbd701e52035\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") "
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.816206    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpk8v\" (UniqueName: \"kubernetes.io/projected/68ff50fc-c751-42f3-8d84-cbd701e52035-kube-api-access-kpk8v\") pod \"68ff50fc-c751-42f3-8d84-cbd701e52035\" (UID: \"68ff50fc-c751-42f3-8d84-cbd701e52035\") "
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.816446    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.819167    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-utilities" (OuterVolumeSpecName: "utilities") pod "68ff50fc-c751-42f3-8d84-cbd701e52035" (UID: "68ff50fc-c751-42f3-8d84-cbd701e52035"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.825656    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ff50fc-c751-42f3-8d84-cbd701e52035-kube-api-access-kpk8v" (OuterVolumeSpecName: "kube-api-access-kpk8v") pod "68ff50fc-c751-42f3-8d84-cbd701e52035" (UID: "68ff50fc-c751-42f3-8d84-cbd701e52035"). InnerVolumeSpecName "kube-api-access-kpk8v". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.918468    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpk8v\" (UniqueName: \"kubernetes.io/projected/68ff50fc-c751-42f3-8d84-cbd701e52035-kube-api-access-kpk8v\") on node \"crc\" DevicePath \"\""
Jan 22 10:16:15 crc kubenswrapper[4824]: I0122 10:16:15.919083    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ff50fc-c751-42f3-8d84-cbd701e52035-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.085985    4824 generic.go:334] "Generic (PLEG): container finished" podID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerID="9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549" exitCode=0
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.086043    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfsgl" event={"ID":"68ff50fc-c751-42f3-8d84-cbd701e52035","Type":"ContainerDied","Data":"9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549"}
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.086118    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nfsgl"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.086759    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nfsgl" event={"ID":"68ff50fc-c751-42f3-8d84-cbd701e52035","Type":"ContainerDied","Data":"85d87d444edd0b5c8f42d087aeb9cf596c87d5eb1dd476e1635f70c4ea369031"}
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.086775    4824 scope.go:117] "RemoveContainer" containerID="9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.089259    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerStarted","Data":"185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91"}
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.135720    4824 scope.go:117] "RemoveContainer" containerID="ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.150417    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nfsgl"]
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.157591    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nfsgl"]
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.163324    4824 scope.go:117] "RemoveContainer" containerID="9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.179959    4824 scope.go:117] "RemoveContainer" containerID="9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549"
Jan 22 10:16:16 crc kubenswrapper[4824]: E0122 10:16:16.180500    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549\": container with ID starting with 9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549 not found: ID does not exist" containerID="9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.180533    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549"} err="failed to get container status \"9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549\": rpc error: code = NotFound desc = could not find container \"9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549\": container with ID starting with 9e8dbbf80f7aa7239d172805dcaba2c76980750b53016fe70b3e37977321b549 not found: ID does not exist"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.180561    4824 scope.go:117] "RemoveContainer" containerID="ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b"
Jan 22 10:16:16 crc kubenswrapper[4824]: E0122 10:16:16.180933    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b\": container with ID starting with ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b not found: ID does not exist" containerID="ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.180960    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b"} err="failed to get container status \"ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b\": rpc error: code = NotFound desc = could not find container \"ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b\": container with ID starting with ee30042cc0fde8e76122b10e5619557f085e4a08863a7f6dce263c4c2b321e9b not found: ID does not exist"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.180975    4824 scope.go:117] "RemoveContainer" containerID="9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801"
Jan 22 10:16:16 crc kubenswrapper[4824]: E0122 10:16:16.181345    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801\": container with ID starting with 9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801 not found: ID does not exist" containerID="9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801"
Jan 22 10:16:16 crc kubenswrapper[4824]: I0122 10:16:16.181406    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801"} err="failed to get container status \"9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801\": rpc error: code = NotFound desc = could not find container \"9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801\": container with ID starting with 9c880f1242263861ec8418d14603b36163be84543baf6bc777b490474f78d801 not found: ID does not exist"
Jan 22 10:16:17 crc kubenswrapper[4824]: I0122 10:16:17.101164    4824 generic.go:334] "Generic (PLEG): container finished" podID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerID="185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91" exitCode=0
Jan 22 10:16:17 crc kubenswrapper[4824]: I0122 10:16:17.101209    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerDied","Data":"185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91"}
Jan 22 10:16:17 crc kubenswrapper[4824]: I0122 10:16:17.413410    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" path="/var/lib/kubelet/pods/68ff50fc-c751-42f3-8d84-cbd701e52035/volumes"
Jan 22 10:16:23 crc kubenswrapper[4824]: I0122 10:16:23.173262    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerStarted","Data":"8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74"}
Jan 22 10:16:23 crc kubenswrapper[4824]: I0122 10:16:23.193679    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dqhv9" podStartSLOduration=3.680514131 podStartE2EDuration="11.193657384s" podCreationTimestamp="2026-01-22 10:16:12 +0000 UTC" firstStartedPulling="2026-01-22 10:16:15.079956649 +0000 UTC m=+2007.983433652" lastFinishedPulling="2026-01-22 10:16:22.593099912 +0000 UTC m=+2015.496576905" observedRunningTime="2026-01-22 10:16:23.192313656 +0000 UTC m=+2016.095790649" watchObservedRunningTime="2026-01-22 10:16:23.193657384 +0000 UTC m=+2016.097134377"
Jan 22 10:16:33 crc kubenswrapper[4824]: I0122 10:16:33.133607    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:33 crc kubenswrapper[4824]: I0122 10:16:33.135512    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:33 crc kubenswrapper[4824]: I0122 10:16:33.198439    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:33 crc kubenswrapper[4824]: I0122 10:16:33.304651    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:33 crc kubenswrapper[4824]: I0122 10:16:33.442901    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqhv9"]
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.264469    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dqhv9" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="registry-server" containerID="cri-o://8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74" gracePeriod=2
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.726842    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.901625    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-catalog-content\") pod \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") "
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.901674    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-utilities\") pod \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") "
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.901857    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4q2q\" (UniqueName: \"kubernetes.io/projected/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-kube-api-access-k4q2q\") pod \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\" (UID: \"2ae351ea-0d5c-4729-aee7-b0ba7adedd01\") "
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.903763    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-utilities" (OuterVolumeSpecName: "utilities") pod "2ae351ea-0d5c-4729-aee7-b0ba7adedd01" (UID: "2ae351ea-0d5c-4729-aee7-b0ba7adedd01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:16:35 crc kubenswrapper[4824]: I0122 10:16:35.908108    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-kube-api-access-k4q2q" (OuterVolumeSpecName: "kube-api-access-k4q2q") pod "2ae351ea-0d5c-4729-aee7-b0ba7adedd01" (UID: "2ae351ea-0d5c-4729-aee7-b0ba7adedd01"). InnerVolumeSpecName "kube-api-access-k4q2q". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.004526    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4q2q\" (UniqueName: \"kubernetes.io/projected/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-kube-api-access-k4q2q\") on node \"crc\" DevicePath \"\""
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.004566    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.059788    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ae351ea-0d5c-4729-aee7-b0ba7adedd01" (UID: "2ae351ea-0d5c-4729-aee7-b0ba7adedd01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.106277    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ae351ea-0d5c-4729-aee7-b0ba7adedd01-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.277119    4824 generic.go:334] "Generic (PLEG): container finished" podID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerID="8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74" exitCode=0
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.277178    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerDied","Data":"8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74"}
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.277221    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dqhv9" event={"ID":"2ae351ea-0d5c-4729-aee7-b0ba7adedd01","Type":"ContainerDied","Data":"81c5ee8a3f4444d210e186e4b1e7027f2bdf6561db5286fd13f75c8729283060"}
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.277259    4824 scope.go:117] "RemoveContainer" containerID="8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.277557    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dqhv9"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.325350    4824 scope.go:117] "RemoveContainer" containerID="185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.338346    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dqhv9"]
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.348934    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dqhv9"]
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.355123    4824 scope.go:117] "RemoveContainer" containerID="da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.386309    4824 scope.go:117] "RemoveContainer" containerID="8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74"
Jan 22 10:16:36 crc kubenswrapper[4824]: E0122 10:16:36.387004    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74\": container with ID starting with 8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74 not found: ID does not exist" containerID="8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.387054    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74"} err="failed to get container status \"8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74\": rpc error: code = NotFound desc = could not find container \"8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74\": container with ID starting with 8ef3c8aaa96d9900e540e5bf87e546f662bc45f7bc74ace1f6c5fe2d61564f74 not found: ID does not exist"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.387084    4824 scope.go:117] "RemoveContainer" containerID="185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91"
Jan 22 10:16:36 crc kubenswrapper[4824]: E0122 10:16:36.387480    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91\": container with ID starting with 185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91 not found: ID does not exist" containerID="185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.387527    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91"} err="failed to get container status \"185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91\": rpc error: code = NotFound desc = could not find container \"185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91\": container with ID starting with 185dbb051c33213ef5d06e9bbaf570f6a0442458944cb2f728b1480dabaebd91 not found: ID does not exist"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.387559    4824 scope.go:117] "RemoveContainer" containerID="da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0"
Jan 22 10:16:36 crc kubenswrapper[4824]: E0122 10:16:36.387984    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0\": container with ID starting with da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0 not found: ID does not exist" containerID="da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0"
Jan 22 10:16:36 crc kubenswrapper[4824]: I0122 10:16:36.388011    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0"} err="failed to get container status \"da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0\": rpc error: code = NotFound desc = could not find container \"da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0\": container with ID starting with da59f88a721b54395f48248ef3c91752aa522323e4d949a4c71b546422b551b0 not found: ID does not exist"
Jan 22 10:16:37 crc kubenswrapper[4824]: I0122 10:16:37.418765    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" path="/var/lib/kubelet/pods/2ae351ea-0d5c-4729-aee7-b0ba7adedd01/volumes"
Jan 22 10:18:00 crc kubenswrapper[4824]: I0122 10:18:00.332701    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:18:00 crc kubenswrapper[4824]: I0122 10:18:00.333186    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:18:30 crc kubenswrapper[4824]: I0122 10:18:30.334185    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:18:30 crc kubenswrapper[4824]: I0122 10:18:30.334866    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.073524    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v9zqb"]
Jan 22 10:18:48 crc kubenswrapper[4824]: E0122 10:18:48.074764    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="registry-server"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.074795    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="registry-server"
Jan 22 10:18:48 crc kubenswrapper[4824]: E0122 10:18:48.074828    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="extract-utilities"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.074846    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="extract-utilities"
Jan 22 10:18:48 crc kubenswrapper[4824]: E0122 10:18:48.074878    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="registry-server"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.074897    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="registry-server"
Jan 22 10:18:48 crc kubenswrapper[4824]: E0122 10:18:48.074926    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="extract-content"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.074942    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="extract-content"
Jan 22 10:18:48 crc kubenswrapper[4824]: E0122 10:18:48.074966    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="extract-content"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.074983    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="extract-content"
Jan 22 10:18:48 crc kubenswrapper[4824]: E0122 10:18:48.075005    4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="extract-utilities"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.075022    4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="extract-utilities"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.075401    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ff50fc-c751-42f3-8d84-cbd701e52035" containerName="registry-server"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.075453    4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae351ea-0d5c-4729-aee7-b0ba7adedd01" containerName="registry-server"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.077782    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.081462    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9zqb"]
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.161178    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-catalog-content\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.161338    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-utilities\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.161394    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg87s\" (UniqueName: \"kubernetes.io/projected/e6b7c398-cc27-42e2-aae1-802b47219dfb-kube-api-access-xg87s\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.259167    4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fhwxz"]
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.260753    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.263066    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-utilities\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.263136    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg87s\" (UniqueName: \"kubernetes.io/projected/e6b7c398-cc27-42e2-aae1-802b47219dfb-kube-api-access-xg87s\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.263232    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-catalog-content\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.263678    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-utilities\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.263779    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-catalog-content\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.279992    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fhwxz"]
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.300286    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg87s\" (UniqueName: \"kubernetes.io/projected/e6b7c398-cc27-42e2-aae1-802b47219dfb-kube-api-access-xg87s\") pod \"redhat-marketplace-v9zqb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") " pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.365095    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-utilities\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.365413    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-catalog-content\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.365589    4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg7gj\" (UniqueName: \"kubernetes.io/projected/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-kube-api-access-kg7gj\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.403869    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.467366    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-utilities\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.467470    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-catalog-content\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.467566    4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg7gj\" (UniqueName: \"kubernetes.io/projected/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-kube-api-access-kg7gj\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.467865    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-utilities\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.468141    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-catalog-content\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.491020    4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg7gj\" (UniqueName: \"kubernetes.io/projected/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-kube-api-access-kg7gj\") pod \"community-operators-fhwxz\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") " pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.586905    4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:48 crc kubenswrapper[4824]: I0122 10:18:48.903772    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9zqb"]
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.085577    4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fhwxz"]
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.311737    4824 generic.go:334] "Generic (PLEG): container finished" podID="e6b7c398-cc27-42e2-aae1-802b47219dfb" containerID="92411ee2e7980a0c47d085c6fb710f0193b6097bb936a4ea2825416985c54b2c" exitCode=0
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.311946    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9zqb" event={"ID":"e6b7c398-cc27-42e2-aae1-802b47219dfb","Type":"ContainerDied","Data":"92411ee2e7980a0c47d085c6fb710f0193b6097bb936a4ea2825416985c54b2c"}
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.312077    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9zqb" event={"ID":"e6b7c398-cc27-42e2-aae1-802b47219dfb","Type":"ContainerStarted","Data":"f47a95c3e239629b655479e0849e31f1cd838a1ae7c15d4a5b605a7c41729da2"}
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.313944    4824 generic.go:334] "Generic (PLEG): container finished" podID="d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" containerID="9c31338195c6c9df62cbab6177dccdef09ad251ece0ec51d1c9848bd4fb74ba3" exitCode=0
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.314138    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhwxz" event={"ID":"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1","Type":"ContainerDied","Data":"9c31338195c6c9df62cbab6177dccdef09ad251ece0ec51d1c9848bd4fb74ba3"}
Jan 22 10:18:49 crc kubenswrapper[4824]: I0122 10:18:49.314886    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhwxz" event={"ID":"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1","Type":"ContainerStarted","Data":"208fc30230225d3d4728a837f83ff91d1bf9f4e32b53c72449bdc930d9e45ff2"}
Jan 22 10:18:51 crc kubenswrapper[4824]: I0122 10:18:51.333565    4824 generic.go:334] "Generic (PLEG): container finished" podID="d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" containerID="640cfaecd672d5775be5709c627caddf58912393096bcdee9690975d218773db" exitCode=0
Jan 22 10:18:51 crc kubenswrapper[4824]: I0122 10:18:51.333920    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhwxz" event={"ID":"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1","Type":"ContainerDied","Data":"640cfaecd672d5775be5709c627caddf58912393096bcdee9690975d218773db"}
Jan 22 10:18:51 crc kubenswrapper[4824]: I0122 10:18:51.337651    4824 generic.go:334] "Generic (PLEG): container finished" podID="e6b7c398-cc27-42e2-aae1-802b47219dfb" containerID="7d8c9c2b328413c92130312ddd856564f37c7084ff1b1466864cd09cce92eed4" exitCode=0
Jan 22 10:18:51 crc kubenswrapper[4824]: I0122 10:18:51.337677    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9zqb" event={"ID":"e6b7c398-cc27-42e2-aae1-802b47219dfb","Type":"ContainerDied","Data":"7d8c9c2b328413c92130312ddd856564f37c7084ff1b1466864cd09cce92eed4"}
Jan 22 10:18:52 crc kubenswrapper[4824]: I0122 10:18:52.363159    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhwxz" event={"ID":"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1","Type":"ContainerStarted","Data":"4f84e681860e9e8bfaddae469034e0ece443a28ed2e17e972c6b68cee66cb893"}
Jan 22 10:18:52 crc kubenswrapper[4824]: I0122 10:18:52.368965    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9zqb" event={"ID":"e6b7c398-cc27-42e2-aae1-802b47219dfb","Type":"ContainerStarted","Data":"fa78c5c5506e99034081e6224abfa60f6b6fb2be1def6cec19b86d62773b18f9"}
Jan 22 10:18:52 crc kubenswrapper[4824]: I0122 10:18:52.384847    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fhwxz" podStartSLOduration=1.614248315 podStartE2EDuration="4.384826768s" podCreationTimestamp="2026-01-22 10:18:48 +0000 UTC" firstStartedPulling="2026-01-22 10:18:49.315400812 +0000 UTC m=+2162.218877805" lastFinishedPulling="2026-01-22 10:18:52.085979265 +0000 UTC m=+2164.989456258" observedRunningTime="2026-01-22 10:18:52.383458042 +0000 UTC m=+2165.286935035" watchObservedRunningTime="2026-01-22 10:18:52.384826768 +0000 UTC m=+2165.288303761"
Jan 22 10:18:52 crc kubenswrapper[4824]: I0122 10:18:52.408396    4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v9zqb" podStartSLOduration=1.6775497499999998 podStartE2EDuration="4.408371397s" podCreationTimestamp="2026-01-22 10:18:48 +0000 UTC" firstStartedPulling="2026-01-22 10:18:49.31341735 +0000 UTC m=+2162.216894343" lastFinishedPulling="2026-01-22 10:18:52.044238997 +0000 UTC m=+2164.947715990" observedRunningTime="2026-01-22 10:18:52.399754021 +0000 UTC m=+2165.303231014" watchObservedRunningTime="2026-01-22 10:18:52.408371397 +0000 UTC m=+2165.311848400"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.404333    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.404754    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.446892    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.491787    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.587346    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.587453    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:58 crc kubenswrapper[4824]: I0122 10:18:58.643210    4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:59 crc kubenswrapper[4824]: I0122 10:18:59.498681    4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:18:59 crc kubenswrapper[4824]: I0122 10:18:59.653975    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9zqb"]
Jan 22 10:19:00 crc kubenswrapper[4824]: I0122 10:19:00.333105    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:19:00 crc kubenswrapper[4824]: I0122 10:19:00.333170    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:19:00 crc kubenswrapper[4824]: I0122 10:19:00.333213    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 10:19:00 crc kubenswrapper[4824]: I0122 10:19:00.333793    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2eaeaf473f7555a3f6d348915a2aedbf1a90414aeb0df2817d91de3ec90015a3"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 10:19:00 crc kubenswrapper[4824]: I0122 10:19:00.333852    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://2eaeaf473f7555a3f6d348915a2aedbf1a90414aeb0df2817d91de3ec90015a3" gracePeriod=600
Jan 22 10:19:00 crc kubenswrapper[4824]: I0122 10:19:00.451885    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v9zqb" podUID="e6b7c398-cc27-42e2-aae1-802b47219dfb" containerName="registry-server" containerID="cri-o://fa78c5c5506e99034081e6224abfa60f6b6fb2be1def6cec19b86d62773b18f9" gracePeriod=2
Jan 22 10:19:01 crc kubenswrapper[4824]: I0122 10:19:01.050112    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fhwxz"]
Jan 22 10:19:01 crc kubenswrapper[4824]: I0122 10:19:01.458958    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fhwxz" podUID="d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" containerName="registry-server" containerID="cri-o://4f84e681860e9e8bfaddae469034e0ece443a28ed2e17e972c6b68cee66cb893" gracePeriod=2
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.468900    4824 generic.go:334] "Generic (PLEG): container finished" podID="d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" containerID="4f84e681860e9e8bfaddae469034e0ece443a28ed2e17e972c6b68cee66cb893" exitCode=0
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.469046    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhwxz" event={"ID":"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1","Type":"ContainerDied","Data":"4f84e681860e9e8bfaddae469034e0ece443a28ed2e17e972c6b68cee66cb893"}
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.485293    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="2eaeaf473f7555a3f6d348915a2aedbf1a90414aeb0df2817d91de3ec90015a3" exitCode=0
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.485502    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"2eaeaf473f7555a3f6d348915a2aedbf1a90414aeb0df2817d91de3ec90015a3"}
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.485537    4824 scope.go:117] "RemoveContainer" containerID="cbdd47c9b3a1c2af08d8a313136db9f7e9c3a7e4c707ddde57f267daf59d2d15"
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.490000    4824 generic.go:334] "Generic (PLEG): container finished" podID="e6b7c398-cc27-42e2-aae1-802b47219dfb" containerID="fa78c5c5506e99034081e6224abfa60f6b6fb2be1def6cec19b86d62773b18f9" exitCode=0
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.490041    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9zqb" event={"ID":"e6b7c398-cc27-42e2-aae1-802b47219dfb","Type":"ContainerDied","Data":"fa78c5c5506e99034081e6224abfa60f6b6fb2be1def6cec19b86d62773b18f9"}
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.898861    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:19:02 crc kubenswrapper[4824]: I0122 10:19:02.900055    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.093460    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg7gj\" (UniqueName: \"kubernetes.io/projected/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-kube-api-access-kg7gj\") pod \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") "
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.093542    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-utilities\") pod \"e6b7c398-cc27-42e2-aae1-802b47219dfb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") "
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.093572    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-catalog-content\") pod \"e6b7c398-cc27-42e2-aae1-802b47219dfb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") "
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.093648    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-catalog-content\") pod \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") "
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.093673    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg87s\" (UniqueName: \"kubernetes.io/projected/e6b7c398-cc27-42e2-aae1-802b47219dfb-kube-api-access-xg87s\") pod \"e6b7c398-cc27-42e2-aae1-802b47219dfb\" (UID: \"e6b7c398-cc27-42e2-aae1-802b47219dfb\") "
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.093702    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-utilities\") pod \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\" (UID: \"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1\") "
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.095781    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-utilities" (OuterVolumeSpecName: "utilities") pod "e6b7c398-cc27-42e2-aae1-802b47219dfb" (UID: "e6b7c398-cc27-42e2-aae1-802b47219dfb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.096961    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-utilities" (OuterVolumeSpecName: "utilities") pod "d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" (UID: "d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.097461    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.100520    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6b7c398-cc27-42e2-aae1-802b47219dfb-kube-api-access-xg87s" (OuterVolumeSpecName: "kube-api-access-xg87s") pod "e6b7c398-cc27-42e2-aae1-802b47219dfb" (UID: "e6b7c398-cc27-42e2-aae1-802b47219dfb"). InnerVolumeSpecName "kube-api-access-xg87s". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.100668    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-kube-api-access-kg7gj" (OuterVolumeSpecName: "kube-api-access-kg7gj") pod "d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" (UID: "d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1"). InnerVolumeSpecName "kube-api-access-kg7gj". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.119872    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6b7c398-cc27-42e2-aae1-802b47219dfb" (UID: "e6b7c398-cc27-42e2-aae1-802b47219dfb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.154904    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" (UID: "d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.199225    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.199267    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg87s\" (UniqueName: \"kubernetes.io/projected/e6b7c398-cc27-42e2-aae1-802b47219dfb-kube-api-access-xg87s\") on node \"crc\" DevicePath \"\""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.199284    4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-utilities\") on node \"crc\" DevicePath \"\""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.199297    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg7gj\" (UniqueName: \"kubernetes.io/projected/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1-kube-api-access-kg7gj\") on node \"crc\" DevicePath \"\""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.199310    4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6b7c398-cc27-42e2-aae1-802b47219dfb-catalog-content\") on node \"crc\" DevicePath \"\""
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.498758    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9zqb" event={"ID":"e6b7c398-cc27-42e2-aae1-802b47219dfb","Type":"ContainerDied","Data":"f47a95c3e239629b655479e0849e31f1cd838a1ae7c15d4a5b605a7c41729da2"}
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.498815    4824 scope.go:117] "RemoveContainer" containerID="fa78c5c5506e99034081e6224abfa60f6b6fb2be1def6cec19b86d62773b18f9"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.498926    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9zqb"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.504651    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fhwxz" event={"ID":"d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1","Type":"ContainerDied","Data":"208fc30230225d3d4728a837f83ff91d1bf9f4e32b53c72449bdc930d9e45ff2"}
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.504743    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fhwxz"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.511025    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerStarted","Data":"2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"}
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.527580    4824 scope.go:117] "RemoveContainer" containerID="7d8c9c2b328413c92130312ddd856564f37c7084ff1b1466864cd09cce92eed4"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.527857    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9zqb"]
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.544244    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9zqb"]
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.549710    4824 scope.go:117] "RemoveContainer" containerID="92411ee2e7980a0c47d085c6fb710f0193b6097bb936a4ea2825416985c54b2c"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.557624    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fhwxz"]
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.561003    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fhwxz"]
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.581160    4824 scope.go:117] "RemoveContainer" containerID="4f84e681860e9e8bfaddae469034e0ece443a28ed2e17e972c6b68cee66cb893"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.614576    4824 scope.go:117] "RemoveContainer" containerID="640cfaecd672d5775be5709c627caddf58912393096bcdee9690975d218773db"
Jan 22 10:19:03 crc kubenswrapper[4824]: I0122 10:19:03.636755    4824 scope.go:117] "RemoveContainer" containerID="9c31338195c6c9df62cbab6177dccdef09ad251ece0ec51d1c9848bd4fb74ba3"
Jan 22 10:19:05 crc kubenswrapper[4824]: I0122 10:19:05.421068    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1" path="/var/lib/kubelet/pods/d42d12c2-7f8b-4664-8cc3-7a2f34e2c7e1/volumes"
Jan 22 10:19:05 crc kubenswrapper[4824]: I0122 10:19:05.423633    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6b7c398-cc27-42e2-aae1-802b47219dfb" path="/var/lib/kubelet/pods/e6b7c398-cc27-42e2-aae1-802b47219dfb/volumes"
Jan 22 10:21:30 crc kubenswrapper[4824]: I0122 10:21:30.332955    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:21:30 crc kubenswrapper[4824]: I0122 10:21:30.333549    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:22:00 crc kubenswrapper[4824]: I0122 10:22:00.333742    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:22:00 crc kubenswrapper[4824]: I0122 10:22:00.334650    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:22:30 crc kubenswrapper[4824]: I0122 10:22:30.333063    4824 patch_prober.go:28] interesting pod/machine-config-daemon-4527v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body=
Jan 22 10:22:30 crc kubenswrapper[4824]: I0122 10:22:30.334630    4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused"
Jan 22 10:22:30 crc kubenswrapper[4824]: I0122 10:22:30.334685    4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4527v"
Jan 22 10:22:30 crc kubenswrapper[4824]: I0122 10:22:30.335345    4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"} pod="openshift-machine-config-operator/machine-config-daemon-4527v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted"
Jan 22 10:22:30 crc kubenswrapper[4824]: I0122 10:22:30.335448    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerName="machine-config-daemon" containerID="cri-o://2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b" gracePeriod=600
Jan 22 10:22:30 crc kubenswrapper[4824]: E0122 10:22:30.557047    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:22:31 crc kubenswrapper[4824]: I0122 10:22:31.358882    4824 generic.go:334] "Generic (PLEG): container finished" podID="e22c2c99-5647-4fc8-9550-7e2a82584d55" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b" exitCode=0
Jan 22 10:22:31 crc kubenswrapper[4824]: I0122 10:22:31.358953    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4527v" event={"ID":"e22c2c99-5647-4fc8-9550-7e2a82584d55","Type":"ContainerDied","Data":"2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"}
Jan 22 10:22:31 crc kubenswrapper[4824]: I0122 10:22:31.359192    4824 scope.go:117] "RemoveContainer" containerID="2eaeaf473f7555a3f6d348915a2aedbf1a90414aeb0df2817d91de3ec90015a3"
Jan 22 10:22:31 crc kubenswrapper[4824]: I0122 10:22:31.359642    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:22:31 crc kubenswrapper[4824]: E0122 10:22:31.359902    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:22:46 crc kubenswrapper[4824]: I0122 10:22:46.404854    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:22:46 crc kubenswrapper[4824]: E0122 10:22:46.406420    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:23:01 crc kubenswrapper[4824]: I0122 10:23:01.404343    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:23:01 crc kubenswrapper[4824]: E0122 10:23:01.405126    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:23:15 crc kubenswrapper[4824]: I0122 10:23:15.404891    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:23:15 crc kubenswrapper[4824]: E0122 10:23:15.411983    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:23:28 crc kubenswrapper[4824]: I0122 10:23:28.403689    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:23:28 crc kubenswrapper[4824]: E0122 10:23:28.404388    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:23:41 crc kubenswrapper[4824]: I0122 10:23:41.409503    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:23:41 crc kubenswrapper[4824]: E0122 10:23:41.410612    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:23:55 crc kubenswrapper[4824]: I0122 10:23:55.403908    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:23:55 crc kubenswrapper[4824]: E0122 10:23:55.404654    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:24:06 crc kubenswrapper[4824]: I0122 10:24:06.407327    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:24:06 crc kubenswrapper[4824]: E0122 10:24:06.408094    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:24:21 crc kubenswrapper[4824]: I0122 10:24:21.405276    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:24:21 crc kubenswrapper[4824]: E0122 10:24:21.406660    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:24:34 crc kubenswrapper[4824]: I0122 10:24:34.404145    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:24:34 crc kubenswrapper[4824]: E0122 10:24:34.405161    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:24:47 crc kubenswrapper[4824]: I0122 10:24:47.414954    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:24:47 crc kubenswrapper[4824]: E0122 10:24:47.415799    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:25:00 crc kubenswrapper[4824]: I0122 10:25:00.404223    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:25:00 crc kubenswrapper[4824]: E0122 10:25:00.404964    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:25:09 crc kubenswrapper[4824]: I0122 10:25:09.497298    4824 generic.go:334] "Generic (PLEG): container finished" podID="39587411-6600-4d7d-9c5a-dc98e8b5c107" containerID="c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce" exitCode=0
Jan 22 10:25:09 crc kubenswrapper[4824]: I0122 10:25:09.497350    4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nnxxz/must-gather-t4pqm" event={"ID":"39587411-6600-4d7d-9c5a-dc98e8b5c107","Type":"ContainerDied","Data":"c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"}
Jan 22 10:25:09 crc kubenswrapper[4824]: I0122 10:25:09.498060    4824 scope.go:117] "RemoveContainer" containerID="c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"
Jan 22 10:25:09 crc kubenswrapper[4824]: I0122 10:25:09.639111    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nnxxz_must-gather-t4pqm_39587411-6600-4d7d-9c5a-dc98e8b5c107/gather/0.log"
Jan 22 10:25:15 crc kubenswrapper[4824]: I0122 10:25:15.410140    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:25:15 crc kubenswrapper[4824]: E0122 10:25:15.411118    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.177959    4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nnxxz/must-gather-t4pqm"]
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.178825    4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-nnxxz/must-gather-t4pqm" podUID="39587411-6600-4d7d-9c5a-dc98e8b5c107" containerName="copy" containerID="cri-o://f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3" gracePeriod=2
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.185601    4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nnxxz/must-gather-t4pqm"]
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.525795    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nnxxz_must-gather-t4pqm_39587411-6600-4d7d-9c5a-dc98e8b5c107/copy/0.log"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.526631    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.542481    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/39587411-6600-4d7d-9c5a-dc98e8b5c107-must-gather-output\") pod \"39587411-6600-4d7d-9c5a-dc98e8b5c107\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") "
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.542541    4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzwj2\" (UniqueName: \"kubernetes.io/projected/39587411-6600-4d7d-9c5a-dc98e8b5c107-kube-api-access-kzwj2\") pod \"39587411-6600-4d7d-9c5a-dc98e8b5c107\" (UID: \"39587411-6600-4d7d-9c5a-dc98e8b5c107\") "
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.547479    4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nnxxz_must-gather-t4pqm_39587411-6600-4d7d-9c5a-dc98e8b5c107/copy/0.log"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.548094    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39587411-6600-4d7d-9c5a-dc98e8b5c107-kube-api-access-kzwj2" (OuterVolumeSpecName: "kube-api-access-kzwj2") pod "39587411-6600-4d7d-9c5a-dc98e8b5c107" (UID: "39587411-6600-4d7d-9c5a-dc98e8b5c107"). InnerVolumeSpecName "kube-api-access-kzwj2". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.548189    4824 generic.go:334] "Generic (PLEG): container finished" podID="39587411-6600-4d7d-9c5a-dc98e8b5c107" containerID="f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3" exitCode=143
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.548239    4824 scope.go:117] "RemoveContainer" containerID="f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.548270    4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nnxxz/must-gather-t4pqm"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.584954    4824 scope.go:117] "RemoveContainer" containerID="c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.650451    4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzwj2\" (UniqueName: \"kubernetes.io/projected/39587411-6600-4d7d-9c5a-dc98e8b5c107-kube-api-access-kzwj2\") on node \"crc\" DevicePath \"\""
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.653798    4824 scope.go:117] "RemoveContainer" containerID="f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3"
Jan 22 10:25:16 crc kubenswrapper[4824]: E0122 10:25:16.654303    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3\": container with ID starting with f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3 not found: ID does not exist" containerID="f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.654381    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3"} err="failed to get container status \"f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3\": rpc error: code = NotFound desc = could not find container \"f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3\": container with ID starting with f2dcdf079562c08f806ae35c756daa830ca1d433128ad6c72ef48426be50ead3 not found: ID does not exist"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.654402    4824 scope.go:117] "RemoveContainer" containerID="c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"
Jan 22 10:25:16 crc kubenswrapper[4824]: E0122 10:25:16.654798    4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce\": container with ID starting with c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce not found: ID does not exist" containerID="c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.654824    4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce"} err="failed to get container status \"c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce\": rpc error: code = NotFound desc = could not find container \"c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce\": container with ID starting with c56b56098334fdacd80e8d8cb1e967770d1da528cbe86dfae71480680cca38ce not found: ID does not exist"
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.663122    4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39587411-6600-4d7d-9c5a-dc98e8b5c107-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "39587411-6600-4d7d-9c5a-dc98e8b5c107" (UID: "39587411-6600-4d7d-9c5a-dc98e8b5c107"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue ""
Jan 22 10:25:16 crc kubenswrapper[4824]: I0122 10:25:16.752287    4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/39587411-6600-4d7d-9c5a-dc98e8b5c107-must-gather-output\") on node \"crc\" DevicePath \"\""
Jan 22 10:25:17 crc kubenswrapper[4824]: I0122 10:25:17.414017    4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39587411-6600-4d7d-9c5a-dc98e8b5c107" path="/var/lib/kubelet/pods/39587411-6600-4d7d-9c5a-dc98e8b5c107/volumes"
Jan 22 10:25:27 crc kubenswrapper[4824]: I0122 10:25:27.407160    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:25:27 crc kubenswrapper[4824]: E0122 10:25:27.409107    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:25:41 crc kubenswrapper[4824]: I0122 10:25:41.404868    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:25:41 crc kubenswrapper[4824]: E0122 10:25:41.405793    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
Jan 22 10:25:55 crc kubenswrapper[4824]: I0122 10:25:55.404190    4824 scope.go:117] "RemoveContainer" containerID="2f41ee252c767456afbf2cbc9e8fc2f561014f2ea07cc91092345dc0a778b95b"
Jan 22 10:25:55 crc kubenswrapper[4824]: E0122 10:25:55.405404    4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4527v_openshift-machine-config-operator(e22c2c99-5647-4fc8-9550-7e2a82584d55)\"" pod="openshift-machine-config-operator/machine-config-daemon-4527v" podUID="e22c2c99-5647-4fc8-9550-7e2a82584d55"
var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515134375507024457 0ustar  coreroot
 Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015134375510017366 5ustar  corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015134367774016526 5ustar  corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015134367774015476 5ustar  corecore